How to process tab-separated files in Spark? -


i have file tab separated. third column should key , entire record should value (as per map reduce concept).

val ceffile = sc.textfile("c:\\text1.txt") val cefdim1 =  ceffile.filter { line => line.startswith("1") } val joinedrdd = ceffile.map(x => x.split("\\t"))  joinedrdd.first().foreach { println }  

i able value of first column not third. can suggest me how accomplish this?

after you've done split x.split("\\t") rdd (which in example called joinedrdd i'm going call parsedrdd since haven't joined yet) going rdd of arrays. turn array of key/value tuples doing parsedrdd.map(r => (r(2), r)). being said - aren't limited map & reduce operations in spark possible data structure might better suited. tab separated files, use spark-csv along spark dataframes if fit eventual problem looking solve.


Comments

Popular posts from this blog

c# - Binding a comma separated list to a List<int> in asp.net web api -

Delphi 7 and decode UTF-8 base64 -

html - Is there any way to exclude a single element from the style? (Bootstrap) -