Skip to main content

Google map API + Encoded polyline

Recently I downloaded Subgurim's google map interfacing library for asp.net. It's cool library supporting many origin javascript functions of Google API. And the key point is that you can write a code directly inside server code(well mine is C#) , and you don't need to care server-client side integration. My aim was to display very long and complex trajectories in map. First I drew lines with Subgurim's AddPolyLine function, but it's quite slow , while zooming it takes 3-4 seconds to redraw lines. So by surfing net I found Encoded polylines which is very optimal way to display polylines. Encoded polyline is like a hashed key which stores lines with their different zoom level. Unfortunately Subgurim's library did not include encoded polyline functionality. So I added some manual code:

private void AddEncodedPolyline(string polyLineID, List tracks, string color, string weight)
{
PolylineEncoder enc = new PolylineEncoder();
EncodedPolyline resPol = enc.Encode(tracks);
string A1 = resPol.Points;
A1 = A1.Replace("{", "{{");
A1 = A1.Replace("}", "}}");
A1 = A1.Replace("\\", "\\\\");
string B1 = resPol.Levels;
string zoomFact = resPol.ZoomFactor.ToString();
string numLevels = resPol.NumLevels.ToString();
string js = string.Format(@"var "+polyLineID+" = new GPolyline.fromEncoded({{ color: '" + color + "', weight: " + weight + ",opacity: 1, points: '" + A1 + "', levels: '" + B1 + "', zoomFactor:" + zoomFact + ", numLevels: " + numLevels + "}});{0}.addOverlay(" + polyLineID + ");", gmpa.GMap_Id);
gmpa.addCustomInsideJavascript(js);
}

gmpa is a subgurim's object for interacting with google map api. I used PolylineEncoder class from Gabriel Svennerberg. he translated original polylineEncoder written in java into C#.
Also I used vertexWelder class for welding vertexes. Welding gave 15 times optimization of lines, though the quality of trajectory was same. It removed close located vertexes from trajectory.

Comments

Popular posts from this blog

Three essential things to do while building Hadoop environment

Last year I setup Hadoop environment by using Cloudera manager. (Basically I followed this video tutorial :  http://www.youtube.com/watch?v=CobVqNMiqww ) I used CDH4(cloudera hadoop)  that included HDFS, MapReduce, Hive, ZooKeeper HBase, Flume and other essential components. It also included YARN (MapReduce 2) but it was not stable so I used MapReduce instead. I installed CDH4 on 10 centos nodes, and I set the Flume to collect twitter data, and by using "crontab" I scheduled the indexing the twitter data in Hive. Anyways, I want to share some of my experiences  and challenges that I faced. First, let me give some problem solutions that everyone must had faced while using Hadoop. 1. vm.swappiness warning on hadoop nodes It is easy to get rid of this warning by just simply running this shell command on nodes: >sysctl -w vm.swappiness=0 More details are written on cloudera's site 2. Make sure to synchronize time on all nodes (otherwise it will give error on n

NLP for Uzbek language

    Natural language processing is an essential tool for text mining in data analysis field. In this post, I want to share my approach in developing stemmer for Uzbek language.      Uzbek language is spoken by 27 million people  around the world and there are a lot of textual materials in internet in uzbek language and it is growing. As I was doing my weekend project " FlipUz " (which is news aggregator for Uzbek news sites) I stumbled on a problem of automatic tagging news into different categories. As this requires a good NLP library, I was not able to find one for Uzbek language. That is how I got a motive to develop a stemmer for Uzbek language.       In short,  Stemming  is an algorithm to remove meaningless suffixes at the end, thus showing the core part of the word. For example: rabbits -> rabbit. As Uzbek language is similar to Turkish, I was curious if there is stemmer for Turkish. And I found this: Turkish Stemmer with Snowball.  Their key approach was to u

Solving java.lang.ClassNotFoundException problem while exporting Jar file from Eclipse

While running my compiled jar files on hadoop cluster I faced this error many times: Caused by: java.lang.ClassNotFoundException: org.apache.hadoop.conf.Configuration Usually, I don't use maven and attach libraries manually to the project. The libraries that I use are mostly apache's libraries related to hadoop, hive, hbase and etc.. After successful compilation I export the project as a Jar file. I tried to export in many ways but I faced this ClassNotFoundException every time. So I realized that exporting my project with attached jar libraries is not the optimal solution for the problem. Then I found this solution on StackOverflow http://stackoverflow.com/questions/2096283/including-jars-in-classpath-on-commandline-javac-or-apt It describes the way how to add required (dependency) jar files while running the main jar. It is simple! You just put all jars into the same folder where your main jar file is and Run: >java -cp *:. com.kh.demo.RealtimeTFIDF Note