Apache spark has a notion of a “data frame”. This object is like a PANDAS data frame or a data frame in R.
I wonder if the auto completion problem as todo with the way code is typically written in spark. I.E. By creating chains of function alls? Each function returns a data frame.
hashTagsDF.groupBy("tag").agg({"tag": "count"}).orderBy("count(tag)", ascending=False).show()
Any suggestions would be greatly appreciated. Bellow is my original post to the apache spark list.
Andy
I started using python3 and jupyter in a chrome browser. I seem to be having trouble with data frame code completion. Regular python functions seems to work correctly.
I wonder if I need to import something so the notebook knows about data frames?
Kind regards
Andy