Merge pull request #90 from madefu/master

This commit is contained in:
Yong Mao
2019-05-27 09:42:42 +08:00
committed by GitHub
23 changed files with 124 additions and 1 deletions

View File

@ -31,4 +31,4 @@ y_pred = regressor.predict(X_test)
# regression evaluation
from sklearn.metrics import r2_score
print(r2_score(Y_test,y_pred))
print(r2_score(Y_test, y_pred))

13
Code/KafkaProducer.py Normal file
View File

@ -0,0 +1,13 @@
#!/usr/bin/python
from kafka import KafkaProducer
kafkaHosts=["kafka01.paas.longfor.sit:9092"
,"kafka02.paas.longfor.sit:9092"
,"kafka03.paas.longfor.sit:9092"]
producer = KafkaProducer(bootstrap_servers=kafkaHosts);
for _ in range(20):
producer.send("testapplog_plm-prototype",b"Hello....")
producer.flush();

24
Code/TestKafka.py Normal file
View File

@ -0,0 +1,24 @@
#!/usr/bin/python
from kafka import KafkaConsumer;
kafkaHosts=["kafka01.paas.longfor.sit:9092"
,"kafka02.paas.longfor.sit:9092"
,"kafka03.paas.longfor.sit:9092"]
'''
earliest
当各分区下有已提交的offset时从提交的offset开始消费无提交的offset时从头开始消费
latest
当各分区下有已提交的offset时从提交的offset开始消费无提交的offset时消费新产生的该分区下的数据
none
topic各分区都存在已提交的offset时从offset后开始消费只要有一个分区不存在已提交的offset则抛出异常
'''
consumer = KafkaConsumer(
bootstrap_servers=kafkaHosts,group_id='mdf_group',auto_offset_reset='latest');
consumer.subscribe("testapplog_plm-prototype");
for msg in consumer:
print(msg.value)

View File

@ -0,0 +1,11 @@
Age,Salary
44,72000
27,48000
30,54000
38,61000
40,78000
35,58000
35,52000
48,79000
50,83000
37,67000
1 Age Salary
2 44 72000
3 27 48000
4 30 54000
5 38 61000
6 40 78000
7 35 58000
8 35 52000
9 48 79000
10 50 83000
11 37 67000

6
Code/my/LinerTest.py Normal file
View File

@ -0,0 +1,6 @@
import pandas as pd
import numpy as np
import matplotlib.pyplot as plt
dataset = pd.read_csv('Data_age_salary.csv');
dataset.iloc[:1]