在spark-shell中将数据保存到cassandra:

var data = normalfill.map(line => line.split("\u0005"))

data.map(
 line => (line(0), line(1), line(2))) 
).saveToCassandra(
 "cui", 
 "oper_ios",
 SomeColumns("user_no", "cust_id", "oper_code","oper_time")
)

 

 saveToCassandra方法 当字段类型是counter的时候,默认行为是计数 

 

CREATE TABLE cui.incr(
 name text,
 count counter,
 PRIMARY KEY (name)
)

 

scala> var rdd = sc.parallelize(Array(("cui", 100 )))
rdd: org.apache.spark.rdd.RDD[(String, Int)] = ParallelCollectionRDD[820] at parallelize at <console>:42

scala>  rdd.saveToCassandra("cui","incr", SomeColumns("name","count"))
16/01/21 16:55:35 INFO core.Cluster: New Cassandra host /172.25.1.158:9042 added
……

// name     count

// cui          100

scala> var rdd = sc.parallelize(Array(("cui", 100 )))
rdd: org.apache.spark.rdd.RDD[(String, Int)] = ParallelCollectionRDD[821] at parallelize at <console>:42

scala>  rdd.saveToCassandra("cui","incr", SomeColumns("name","count"))

// name     count

// cui          200

相关文章:

  • 2021-11-11
  • 2021-08-30
  • 2022-01-12
  • 2021-12-15
  • 2022-12-23
  • 2022-12-23
  • 2021-04-18
猜你喜欢
  • 2022-12-23
  • 2021-06-04
  • 2021-12-24
  • 2021-11-27
  • 2022-12-23
  • 2022-12-23
  • 2021-09-02
相关资源
相似解决方案