how to use jni in spark?

Hellen picture Hellen · Jan 14, 2014 · Viewed 7k times · Source

I want to use jni to call my c++ lib in spark. When i sbt run my program, it shows that java.lang.UnsatisfiedLinkError: no hq_Image_Process in java.library.path , so obviously the program can not find my hq_Image_Process.so .

In hadoop, -files can distribute the xxx.so file to the slaves like this:

[hadoop@Master ~]$ hadoop jar JniTest3.jar -files /home/hadoop/Documents/java/jni1/bin/libFakeSegmentForJni.so FakeSegmentForJni.TestFakeSegmentForJni input output

Are there any ways to call my hq_Image_Process.so like hadoop in spark? I would appreciate any help.

Answer

Wildfire picture Wildfire · Jan 15, 2014

First of all, the native library must be preinstalled on all worker nodes. Path to that library must be specified in spark-env.sh:

export SPARK_LIBRARY_PATH=/path/to/native/library

SPARK_PRINT_LAUNCH_COMMAND environment variable might be used to diagnose it:

export SPARK_PRINT_LAUNCH_COMMAND=1

If everything's set correctly, you will see output like this:

Spark Command:
/path/to/java -cp <long list of jars> -Djava.library.path=/path/to/native/library <etc>
========================================