-
Notifications
You must be signed in to change notification settings - Fork 147
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Getting started task #117
Comments
Hi @ELMERWANG
Btw, if you are new to stream mining, I suggest this open book to get started |
Hi @hmgomes, Thank you for your reply, I have tried the command that you provide, however the error now changed to: Error message: And the result file is empty... |
Hi @ELMERWANG That is interesting. Can you try using Spark 2.4.5? Cheers, |
Hi @hmgomes, Thanks, may I ask which version should I download? With or without hadoop? [ ] spark-2.4.5-bin-hadoop2.6.tgz 2020-02-02 20:27 220M [ ] spark-2.4.5-bin-hadoop2.7.tgz 2020-02-02 20:27 222M [ ] spark-2.4.5-bin-without-hadoop-scala-2.12.tgz 2020-02-02 20:27 139M [ ] spark-2.4.5-bin-without-hadoop.tgz 2020-02-02 20:27 160M [ ] spark-2.4.5.tgz 2020-02-02 20:27 15M |
@ELMERWANG Hi,I met the same problem as yours. Do you solve your problem?I want to get some suggestions to solve these problems. Thank you! Infrastructure details When I run these Command lines I get some errors:
after I ran
|
Unfortunately, no luck on fixing it, sorry.. |
Hi @ELMERWANG I have the version with Hadoop Regards, |
Hi @LYZ1125
Regards, |
Hi,
I am quite new to stream mining and I am trying to get the streamDM running. However when I finished setting up all the dependencies I got some issues reported below.
Expected behavior
I want to obtain the confusion matrix by using the example commands.
Observed behavior
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
21/03/09 18:40:50 INFO SparkContext: Running Spark version 2.1.0
21/03/09 18:40:50 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
21/03/09 18:40:50 WARN Utils: Your hostname, localhost.localdomain resolves to a loopback address: 127.0.0.1; using 192.168.122.1 instead (on interface virbr0)
21/03/09 18:40:50 WARN Utils: Set SPARK_LOCAL_IP if you need to bind to another address
21/03/09 18:40:50 INFO SecurityManager: Changing view acls to: shunyao
21/03/09 18:40:50 INFO SecurityManager: Changing modify acls to: shunyao
21/03/09 18:40:50 INFO SecurityManager: Changing view acls groups to:
21/03/09 18:40:50 INFO SecurityManager: Changing modify acls groups to:
21/03/09 18:40:50 INFO SecurityManager: SecurityManager: authentication disabled; ui acls disabled; users with view permissions: Set(shunyao); groups with view permissions: Set(); users with modify permissions: Set(shunyao); groups with modify permissions: Set()
21/03/09 18:40:50 INFO Utils: Successfully started service 'sparkDriver' on port 41123.
21/03/09 18:40:50 INFO SparkEnv: Registering MapOutputTracker
21/03/09 18:40:50 INFO SparkEnv: Registering BlockManagerMaster
21/03/09 18:40:50 INFO BlockManagerMasterEndpoint: Using org.apache.spark.storage.DefaultTopologyMapper for getting topology information
21/03/09 18:40:50 INFO BlockManagerMasterEndpoint: BlockManagerMasterEndpoint up
21/03/09 18:40:50 INFO DiskBlockManager: Created local directory at /tmp/blockmgr-8d18da96-7b2c-48c9-bbd2-b49a29b16274
21/03/09 18:40:50 INFO MemoryStore: MemoryStore started with capacity 413.9 MB
21/03/09 18:40:50 INFO SparkEnv: Registering OutputCommitCoordinator
21/03/09 18:40:51 INFO Utils: Successfully started service 'SparkUI' on port 4040.
21/03/09 18:40:51 INFO SparkUI: Bound SparkUI to 0.0.0.0, and started at http://192.168.122.1:4040
21/03/09 18:40:51 INFO SparkContext: Added JAR file:/home/shunyao/Downloads/streamDM-0.2/scripts/../target/scala-2.10/streamdm-spark-streaming-_2.10-0.2.jar at spark://192.168.122.1:41123/jars/streamdm-spark-streaming-_2.10-0.2.jar with timestamp 1615333251155
21/03/09 18:40:51 INFO Executor: Starting executor ID driver on host localhost
21/03/09 18:40:51 INFO Utils: Successfully started service 'org.apache.spark.network.netty.NettyBlockTransferService' on port 45261.
21/03/09 18:40:51 INFO NettyBlockTransferService: Server created on 192.168.122.1:45261
21/03/09 18:40:51 INFO BlockManager: Using org.apache.spark.storage.RandomBlockReplicationPolicy for block replication policy
21/03/09 18:40:51 INFO BlockManagerMaster: Registering BlockManager BlockManagerId(driver, 192.168.122.1, 45261, None)
21/03/09 18:40:51 INFO BlockManagerMasterEndpoint: Registering block manager 192.168.122.1:45261 with 413.9 MB RAM, BlockManagerId(driver, 192.168.122.1, 45261, None)
21/03/09 18:40:51 INFO BlockManagerMaster: Registered BlockManager BlockManagerId(driver, 192.168.122.1, 45261, None)
21/03/09 18:40:51 INFO BlockManager: Initialized BlockManager: BlockManagerId(driver, 192.168.122.1, 45261, None)
Exception in thread "main" java.lang.Exception: Problem creating instance of class: EvaluatePrequential
at com.github.javacliparser.ClassOption.cliStringToObject(ClassOption.java:139)
at org.apache.spark.streamdm.streamDMJob$.main(streamDMJob.scala:43)
at org.apache.spark.streamdm.streamDMJob.main(streamDMJob.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:738)
at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:187)
at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:212)
at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:126)
at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.lang.IllegalArgumentException: Problems with option: streamReader
at com.github.javacliparser.ClassOption.setValueViaCLIString(ClassOption.java:60)
at com.github.javacliparser.AbstractOption.resetToDefault(AbstractOption.java:90)
at com.github.javacliparser.AbstractClassOption.(AbstractClassOption.java:84)
at com.github.javacliparser.AbstractClassOption.(AbstractClassOption.java:63)
at com.github.javacliparser.ClassOption.(ClassOption.java:34)
at org.apache.spark.streamdm.tasks.EvaluatePrequential.(EvaluatePrequential.scala:47)
at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)
at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:62)
at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
at java.lang.Class.newInstance(Class.java:442)
at com.github.javacliparser.ClassOption.cliStringToObject(ClassOption.java:137)
... 11 more
Caused by: java.lang.Exception: Class not found: FileReader
at com.github.javacliparser.ClassOption.cliStringToObject(ClassOption.java:132)
at com.github.javacliparser.ClassOption.setValueViaCLIString(ClassOption.java:57)
... 22 more
21/03/09 18:40:51 INFO SparkContext: Invoking stop() from shutdown hook
21/03/09 18:40:51 INFO SparkUI: Stopped Spark web UI at http://192.168.122.1:4040
21/03/09 18:40:51 INFO MapOutputTrackerMasterEndpoint: MapOutputTrackerMasterEndpoint stopped!
21/03/09 18:40:51 INFO MemoryStore: MemoryStore cleared
21/03/09 18:40:51 INFO BlockManager: BlockManager stopped
21/03/09 18:40:51 INFO BlockManagerMaster: BlockManagerMaster stopped
21/03/09 18:40:51 INFO OutputCommitCoordinator$OutputCommitCoordinatorEndpoint: OutputCommitCoordinator stopped!
21/03/09 18:40:51 INFO SparkContext: Successfully stopped SparkContext
21/03/09 18:40:51 INFO ShutdownHookManager: Shutdown hook called
21/03/09 18:40:51 INFO ShutdownHookManager: Deleting directory /tmp/spark-6f0a2908-b3f5-477c-bccc-6302be0e18f8
Steps to reproduce the issue
I tried different OS: macOS, Kali, Ubuntu and CentOS 7.
These steps are what I did and followed:
Step 1: Download & install the Java openjdk8.
Step 2: Download & install the Spark (I tried 2.3.2 and 2.1.0 with Hadoop or without Hadoop)
Step 3: Download & install Scala (from https://www.scala-lang.org/files/archive/scala-2.11.8.deb)
Step 4: Download & install SBT (from https://sbt.bintray.com/native-packages/sbt/0.13.15/sbt-0.13.15.tgz)
Step 5: Download StreamDM from Github.
Step 6: Change the SPARK_HOME to the Spark path.
Step 7: In the streamDM directory, run sbt package
Step 8: In the /scripts, run ./spark.sh or ./spark.sh "EvaluatePrequential -l (SGDLearner -l 0.01 -o LogisticLoss -r ZeroRegularizer) –s (FileReader –k 100 –d 60 –f ../data/mydata)"
Result: Errors occurred.
Command line
sbt package
./spark.sh
./spark.sh "EvaluatePrequential -l (SGDLearner -l 0.01 -o LogisticLoss -r ZeroRegularizer) –s (FileReader –k 100 –d 60 –f ../data/mydata)"
Infrastructure details (Contains the versions that I tried)
For the Spark, do I need to download with Hadoop or without Hadoop or the file name only contains spark-x.x.x.tgz?
Such as:
Is anything wrong with my setup? I produced the same error on different OS.
Please help, thanks in advance :)
The text was updated successfully, but these errors were encountered: