File tree Expand file tree Collapse file tree 6 files changed +22
-36
lines changed
main/java/com/nasdaq/ncdsclient Expand file tree Collapse file tree 6 files changed +22
-36
lines changed Original file line number Diff line number Diff line change 7
7
<parent >
8
8
<groupId >com.nasdaq.ncds</groupId >
9
9
<artifactId >ncds</artifactId >
10
- <version >0.8.5 </version >
10
+ <version >0.9.0 </version >
11
11
</parent >
12
12
13
13
<artifactId >ncds-sdk</artifactId >
56
56
<artifactId >slf4j-simple</artifactId >
57
57
<scope >test</scope >
58
58
</dependency >
59
- <!-- Kafka -auth -->
60
- <dependency >
61
- <groupId >io.strimzi</groupId >
62
- <artifactId >kafka-oauth-common</artifactId >
63
- </dependency >
64
-
65
59
<!-- Testing -->
66
60
67
61
<!-- Kafka -->
Original file line number Diff line number Diff line change 7
7
import com .nasdaq .ncdsclient .internal .utils .IsItJunit ;
8
8
import com .nasdaq .ncdsclient .internal .utils .KafkaConfigLoader ;
9
9
import com .nasdaq .ncdsclient .news .NewsUtil ;
10
- import io .strimzi .kafka .oauth .common .ConfigProperties ;
11
10
import org .apache .avro .Schema ;
12
11
import org .apache .kafka .clients .consumer .ConsumerConfig ;
13
12
import org .apache .kafka .clients .consumer .KafkaConsumer ;
@@ -216,14 +215,6 @@ public KafkaConsumer getNewsConsumer(String topic) throws Exception {
216
215
}
217
216
}
218
217
219
- private String getDate (){
220
- // Get Today's EST date
221
- DateFormat dateformat = new SimpleDateFormat ("yyyy-MM-dd" );
222
- dateformat .setTimeZone (TimeZone .getTimeZone ("America/New_York" ));
223
- String date = dateformat .format (new Date ());
224
- return date ;
225
- }
226
-
227
218
private KafkaConsumer seekToMidNight (TopicPartition topicPartition ){
228
219
Map <TopicPartition ,Long > timestmaps = new HashMap ();
229
220
timestmaps .put (topicPartition , getTodayMidNightTimeStamp ());
Original file line number Diff line number Diff line change 2
2
3
3
import com .nasdaq .ncdsclient .internal .utils .IsItJunit ;
4
4
import com .nasdaq .ncdsclient .internal .utils .KafkaConfigLoader ;
5
- import io .strimzi .kafka .oauth .common .ConfigProperties ;
6
5
import org .apache .avro .Schema ;
7
6
import org .apache .avro .generic .GenericRecord ;
8
7
import org .apache .kafka .clients .consumer .*;
@@ -121,9 +120,9 @@ private KafkaAvroConsumer getConsumer(String cleindId) throws Exception {
121
120
//Properties kafkaProps = null;
122
121
try {
123
122
124
- if (!IsItJunit .isJUnitTest ()) {
125
- ConfigProperties .resolveAndExportToSystemProperties (securityProps );
126
- }
123
+ // if(!IsItJunit.isJUnitTest()) {
124
+ // ConfigProperties.resolveAndExportToSystemProperties(securityProps);
125
+ // }
127
126
128
127
Schema .Parser parser = new Schema .Parser ();
129
128
//controlMessageSchema = parser.parse(ClassLoader.getSystemResourceAsStream("ControlMessageSchema.avsc"));
@@ -137,7 +136,7 @@ private KafkaAvroConsumer getConsumer(String cleindId) throws Exception {
137
136
kafkaProps .put (ConsumerConfig .AUTO_OFFSET_RESET_CONFIG , OffsetResetStrategy .EARLIEST .toString ().toLowerCase ());
138
137
kafkaProps .put (ConsumerConfig .GROUP_ID_CONFIG , cleindId );
139
138
kafkaProps .put (ConsumerConfig .MAX_PARTITION_FETCH_BYTES_CONFIG , "5048576" );
140
- ConfigProperties .resolve (kafkaProps );
139
+ // ConfigProperties.resolve(kafkaProps);
141
140
}
142
141
catch (Exception e ) {
143
142
throw e ;
Original file line number Diff line number Diff line change
1
+ <?xml version =" 1.0" encoding =" UTF-8" ?>
2
+ <Configuration status =" DEBUG" >
3
+ <Appenders >
4
+ <File name =" LogToFile" fileName =" logs/unittest/log.out" append =" false" immediateFlush =" true" >
5
+ <PatternLayout >
6
+ <Pattern >%d{yyyy-MM-dd HH:mm:ss} %-5p %c{1}:%L - %m%n</Pattern >
7
+ </PatternLayout >
8
+ </File >
9
+ </Appenders >
10
+ <Loggers >
11
+ <Root level =" INFO" >
12
+ <AppenderRef ref =" LogToFile" />
13
+ </Root >
14
+ </Loggers >
15
+ </Configuration >
Original file line number Diff line number Diff line change 7
7
<parent >
8
8
<groupId >com.nasdaq.ncds</groupId >
9
9
<artifactId >ncds</artifactId >
10
- <version >0.8.5 </version >
10
+ <version >0.9.0 </version >
11
11
</parent >
12
12
13
13
<artifactId >ncdssdk-client</artifactId >
Original file line number Diff line number Diff line change 5
5
<modelVersion >4.0.0</modelVersion >
6
6
<groupId >com.nasdaq.ncds</groupId >
7
7
<artifactId >ncds</artifactId >
8
- <version >0.8.5 </version >
8
+ <version >0.9.0 </version >
9
9
<packaging >pom</packaging >
10
10
<name >Nasdaq Cloud Data Service </name >
11
11
23
23
<log4j .version>2.20.0</log4j .version>
24
24
<netty .version>4.1.94.Final</netty .version>
25
25
<slf4jVersion >1.7.36</slf4jVersion >
26
- <strimzi .oauth.version>0.12.0</strimzi .oauth.version>
27
26
<checkstyle .skip>true</checkstyle .skip>
28
27
<junit5 .version>5.7.2</junit5 .version>
29
28
<junit5PlatformProvider .version>1.3.2</junit5PlatformProvider .version>
116
115
<version >${kafka.version} </version >
117
116
</dependency >
118
117
119
- <!-- Kafka Auth Dependency-->
120
- <dependency >
121
- <groupId >io.strimzi</groupId >
122
- <artifactId >kafka-oauth-client</artifactId >
123
- <version >${strimzi.oauth.version} </version >
124
- </dependency >
125
- <dependency >
126
- <groupId >io.strimzi</groupId >
127
- <artifactId >kafka-oauth-common</artifactId >
128
- <version >${strimzi.oauth.version} </version >
129
- </dependency >
130
-
131
118
<dependency >
132
119
<groupId >org.slf4j</groupId >
133
120
<artifactId >slf4j-log4j12</artifactId >
You can’t perform that action at this time.
0 commit comments