I want to use jni to call my c++ lib in spark. When i sbt run my program, it shows that java.lang.UnsatisfiedLinkError: no hq_Image_Process in java.library.path , so obviously the program can not find my hq_Image_Process.so .
In hadoop, -files can distribute the xxx.so file to the slaves like this:
[hadoop@Master ~]$ hadoop jar JniTest3.jar -files /home/hadoop/Documents/java/jni1/bin/libFakeSegmentForJni.so FakeSegmentForJni.TestFakeSegmentForJni input output
Are there any ways to call my hq_Image_Process.so like hadoop in spark? I would appreciate any help.
First of all, the native library must be preinstalled on all worker nodes. Path to that library must be specified in spark-env.sh
:
export SPARK_LIBRARY_PATH=/path/to/native/library
SPARK_PRINT_LAUNCH_COMMAND
environment variable might be used to diagnose it:
export SPARK_PRINT_LAUNCH_COMMAND=1
If everything's set correctly, you will see output like this:
Spark Command:
/path/to/java -cp <long list of jars> -Djava.library.path=/path/to/native/library <etc>
========================================