Failed to parse private key: Error: Invalid PEM formatted message - firebase

I am trying to deploy a NEXT JS app that uses firebase-admin on Vercel.
import * as firebaseAdmin from 'firebase-admin';
import firebase from 'firebase/app';
if (!firebaseAdmin.apps.length) {
firebaseAdmin.initializeApp({
credential: firebaseAdmin.credential.cert({
privateKey: process.env.NEXT_PUBLIC_FIREBASE_PRIVATE_KEY,
clientEmail: process.env.NEXT_PUBLIC_FIREBASE_CLIENT_EMAIL,
projectId: process.env.NEXT_PUBLIC_FIREBASE_PROJECT_ID,
}),
databaseURL: process.env.NEXT_PUBLIC_FIREBASE_DATABASE_URL,
});
}
When I deploy it to vercel with environment variables as:
NEXT_PUBLIC_FIREBASE_PRIVATE_KEY="-----BEGIN PRIVATE KEY-----\n....\n-----END PRIVATE KEY-----\n"
I get this error message.
> Build error occurred
Error: Certificate object must contain a string "private_key" property.
at FirebaseAppError.FirebaseError [as constructor] (/vercel/path0/node_modules/firebase-admin/lib/utils/error.js:42:28)
at FirebaseAppError.PrefixedFirebaseError [as constructor] (/vercel/path0/node_modules/firebase-admin/lib/utils/error.js:88:28)
at new FirebaseAppError (/vercel/path0/node_modules/firebase-admin/lib/utils/error.js:122:28)
at new Certificate (/vercel/path0/node_modules/firebase-admin/lib/auth/credential.js:118:19)
at new CertCredential (/vercel/path0/node_modules/firebase-admin/lib/auth/credential.js:187:64)
at Object.cert (/vercel/path0/node_modules/firebase-admin/lib/firebase-namespace.js:220:58)
at Object.3996 (/vercel/path0/.next/server/chunks/241.js:115:76)
at __webpack_require__ (/vercel/path0/.next/server/webpack-runtime.js:25:42)
at /vercel/path0/.next/server/pages/dashboard.js:124:81
at Function.__webpack_require__.a (/vercel/path0/.next/server/webpack-runtime.js:103:13) {
type: 'FirebaseAppError',
errorInfo: {
code: 'app/invalid-credential',
message: 'Certificate object must contain a string "private_key" property.'
},
codePrefix: 'app'
}
Error! Command "npm run build" exited with 1
Error: Command "vercel build" exited with 1
However running npm run build builds with no error locally.
.env.local
info - Linting and checking validity of types
info - Creating an optimized production build
info - Compiled successfully
info - Collecting page data
info - Generating static pages (4/4)
info - Finalizing page optimization
Page Size First Load JS
┌ ○ / (466 ms) 461 B 271 kB
├ /_app 0 B 270 kB
├ ○ /404 (450 ms) 526 B 271 kB
├ λ /api/hello 0 B 270 kB
+ First Load JS shared by all 270 kB
├ chunks/framework-a87821de553db91d.js 45 kB
├ chunks/main-567f0ec5ceee81fc.js 29.7 kB
├ chunks/pages/_app-b443b414c9e8f379.js 195 kB
├ chunks/webpack-42cdea76c8170223.js 1.07 kB
└ css/966e875c066cf46b.css 5.18 kB
λ (Server) server-side renders at runtime (uses getInitialProps or getServerSideProps)
○ (Static) automatically rendered as static HTML (uses no initial props)
I do not understand how to fix this because this runs locally and builds without any errors.

I found the answer.
.env.local can have values like
NEXT_PUBLIC_FIREBASE_PRIVATE_KEY="-----BEGIN PRIVATE KEY-----\n....\n-----END PRIVATE KEY-----\n"
But in vercel environment variables add the keys without the double quotes "".
NEXT_PUBLIC_FIREBASE_PRIVATE_KEY=-----BEGIN PRIVATE KEY-----\n....\n-----END PRIVATE KEY-----\n
When you add the keys it should reformat itself and look something like this.
You can console.log the values locally, that will look like this below and paste the exact same value in Vercel env.
-----BEGIN PRIVATE KEY-----
your
secret
key
-----END PRIVATE KEY-----

I ran into this problem too. Removing the quotes didn't help this far.
However, my comment is related to a different thing: You are setting the private key to a NEXT_PUBLIC var (accessible via front-end, hence, exposed).
Just that. Sorry if it's not relevant.

Related

How to solve Firebase Cloud functions deploy error

I'm deploying using Ionic + Capacitor + Firebase.
While I run cli to deploy firebase functions, it having below errors.
The solution I'm having right now is delete the node_modules folder of Ionic project.
Here is my system info:
"firebase": "^9.17.1",
"firebase-functions": "^4.2.1",
`Ionic:
Ionic CLI : 6.20.8 (/usr/local/lib/node_modules/#ionic/cli)
Ionic Framework : #ionic/angular 6.5.2
#angular-devkit/build-angular : 15.1.4
#angular-devkit/schematics : 15.1.4
#angular/cli : 15.1.4
#ionic/angular-toolkit : 7.0.0
Capacitor:
Capacitor CLI : 4.6.3
#capacitor/android : 4.6.3
#capacitor/core : 4.6.3
#capacitor/ios : 4.6.3
Utility:
cordova-res : 0.15.4
native-run : 1.7.1
System:
NodeJS : v18.12.1 (/usr/local/bin/node)
npm : 9.4.0
OS : macOS Monterey`
Error:
`=== Deploying to 'project'...
i deploying functions
Running command: npm --prefix "$RESOURCE_DIR" run lint
lint
tslint --project tsconfig.json
Running command: npm --prefix "$RESOURCE_DIR" run build
build
tsc
../node_modules/#types/jasmine/index.d.ts:25:1 - error TS6200: Definitions of the following identifiers conflict with those in another file: beforeAll, beforeEach, afterAll, afterEach, describe, fdescribe, xdescribe, it, fit, xit, expect, DEFAULT_TIMEOUT_INTERVAL, CustomMatcherFactory, CustomEqualityTester
25 type ImplementationCallback = jasmine.ImplementationCallback;
../node_modules/#types/jest/index.d.ts:33:1
33 declare var beforeAll: jest.Lifecycle;
~~~~~~~
Conflicts are in this file.
../node_modules/#types/jasmine/index.d.ts:405:9 - error TS2374: Duplicate index signature for type 'number'.
405 [n: number]: T;
~~~~~~~~~~~~~~~
../node_modules/#types/jasmine/index.d.ts:408:15 - error TS2428: All declarations of 'ArrayContaining' must have identical type parameters.
408 interface ArrayContaining<T> extends AsymmetricMatcher<any> {
~~~~~~~~~~~~~~~
../node_modules/#types/jasmine/index.d.ts:413:15 - error TS2428: All declarations of 'ObjectContaining' must have identical type parameters.
413 interface ObjectContaining<T> extends AsymmetricMatcher<T> {
~~~~~~~~~~~~~~~~
../node_modules/#types/jasmine/index.d.ts:451:9 - error TS2374: Duplicate index signature for type 'string'.
451 [name: string]: CustomMatcherFactory;
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../node_modules/#types/jasmine/index.d.ts:460:9 - error TS2687: All declarations of 'message' must have identical modifiers.
460 message?: string | undefined;
~~~~~~~
../node_modules/#types/jasmine/index.d.ts:1071:15 - error TS2428: All declarations of 'SpyAnd' must have identical type parameters.
1071 interface SpyAnd<Fn extends Func> {
~~~~~~
../node_modules/#types/jasmine/index.d.ts:1092:15 - error TS2428: All declarations of 'Calls' must have identical type parameters.
1092 interface Calls<Fn extends Func> {
~~~~~
../node_modules/#types/jasmine/index.d.ts:1115:15 - error TS2428: All declarations of 'CallInfo' must have identical type parameters.
1115 interface CallInfo<Fn extends Func> {
~~~~~~~~
../node_modules/#types/jest/index.d.ts:33:1 - error TS6200: Definitions of the following identifiers conflict with those in another file: beforeAll, beforeEach, afterAll, afterEach, describe, fdescribe, xdescribe, it, fit, xit, expect, DEFAULT_TIMEOUT_INTERVAL, CustomMatcherFactory, CustomEqualityTester
33 declare var beforeAll: jest.Lifecycle;
../node_modules/#types/jasmine/index.d.ts:25:1
25 type ImplementationCallback = jasmine.ImplementationCallback;
~~~~
Conflicts are in this file.
../node_modules/#types/jest/index.d.ts:1343:46 - error TS2314: Generic type 'ArrayContaining' requires 1 type argument(s).
1343 function arrayContaining(sample: any[]): ArrayContaining;
~~~~~~~~~~~~~~~
../node_modules/#types/jest/index.d.ts:1344:45 - error TS2314: Generic type 'ObjectContaining' requires 1 type argument(s).
1344 function objectContaining(sample: any): ObjectContaining;
~~~~~~~~~~~~~~~~
../node_modules/#types/jest/index.d.ts:1370:15 - error TS2428: All declarations of 'ArrayContaining' must have identical type parameters.
1370 interface ArrayContaining {
~~~~~~~~~~~~~~~
../node_modules/#types/jest/index.d.ts:1376:15 - error TS2428: All declarations of 'ObjectContaining' must have identical type parameters.
1376 interface ObjectContaining {
~~~~~~~~~~~~~~~~
../node_modules/#types/jest/index.d.ts:1379:9 - error TS2386: Overload signatures must all be optional or required.
1379 jasmineToString(): string;
~~~~~~~~~~~~~~~
../node_modules/#types/jest/index.d.ts:1385:14 - error TS2314: Generic type 'SpyAnd' requires 1 type argument(s).
1385 and: SpyAnd;
~~~~~~
../node_modules/#types/jest/index.d.ts:1386:16 - error TS2314: Generic type 'Calls' requires 1 type argument(s).
1386 calls: Calls;
~~~~~
../node_modules/#types/jest/index.d.ts:1392:15 - error TS2428: All declarations of 'SpyAnd' must have identical type parameters.
1392 interface SpyAnd {
~~~~~~
../node_modules/#types/jest/index.d.ts:1425:15 - error TS2428: All declarations of 'Calls' must have identical type parameters.
1425 interface Calls {
~~~~~
../node_modules/#types/jest/index.d.ts:1451:16 - error TS2314: Generic type 'CallInfo' requires 1 type argument(s).
1451 all(): CallInfo[];
~~~~~~~~
../node_modules/#types/jest/index.d.ts:1456:23 - error TS2314: Generic type 'CallInfo' requires 1 type argument(s).
1456 mostRecent(): CallInfo;
~~~~~~~~
../node_modules/#types/jest/index.d.ts:1461:18 - error TS2314: Generic type 'CallInfo' requires 1 type argument(s).
1461 first(): CallInfo;
~~~~~~~~
../node_modules/#types/jest/index.d.ts:1468:15 - error TS2428: All declarations of 'CallInfo' must have identical type parameters.
1468 interface CallInfo {
~~~~~~~~
../node_modules/#types/jest/index.d.ts:1472:9 - error TS2717: Subsequent property declarations must have the same type. Property 'object' must be of type 'ThisType', but here has type 'any'.
1472 object: any;
~~~~~~
../node_modules/#types/jasmine/index.d.ts:1117:9
1117 object: ThisType;
~~~~~~
'object' was also declared here.
../node_modules/#types/jest/index.d.ts:1476:9 - error TS2717: Subsequent property declarations must have the same type. Property 'args' must be of type 'Parameters', but here has type 'any[]'.
1476 args: any[];
~~~~
../node_modules/#types/jasmine/index.d.ts:1119:9
1119 args: Parameters;
~~~~
'args' was also declared here.
../node_modules/#types/jest/index.d.ts:1480:9 - error TS2717: Subsequent property declarations must have the same type. Property 'returnValue' must be of type 'ReturnType', but here has type 'any'.
1480 returnValue: any;
~~~~~~~~~~~
../node_modules/#types/jasmine/index.d.ts:1121:9
1121 returnValue: ReturnType;
~~~~~~~~~~~
'returnValue' was also declared here.
../node_modules/#types/jest/index.d.ts:1484:9 - error TS2374: Duplicate index signature for type 'string'.
1484 [index: string]: CustomMatcherFactory;
~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~
../node_modules/#types/jest/index.d.ts:1505:9 - error TS2687: All declarations of 'message' must have identical modifiers.
1505 message: string | (() => string);
~~~~~~~
../node_modules/#types/jest/index.d.ts:1505:9 - error TS2717: Subsequent property declarations must have the same type. Property 'message' must be of type 'string | undefined', but here has type 'string | (() => string)'.
1505 message: string | (() => string);
~~~~~~~
../node_modules/#types/jasmine/index.d.ts:460:9
460 message?: string | undefined;
~~~~~~~
'message' was also declared here.
../node_modules/#types/jest/index.d.ts:1510:9 - error TS2374: Duplicate index signature for type 'number'.
1510 [n: number]: T;
~~~~~~~~~~~~~~~
Found 30 errors in 2 files.
Errors Files
9 ../node_modules/#types/jasmine/index.d.ts:25
21 ../node_modules/#types/jest/index.d.ts:33
Error: functions predeploy error: Command terminated with non-zero exit code 2`
Should able to run firebase deploy --only functions directly.

How to re-create an equivalent to linux bash statement in Deno with Deno.Run

How to re-create an equivalent to following Linux bash statement in Deno?
docker compose exec container_name -uroot -ppass db_name < ./dbDump.sql
I have tried the following:
const encoder = new TextEncoder
const p = await Deno.run({
cmd: [
'docker',
'compose',
'exec',
'container_name',
'mysql',
'-uroot',
'-ppass',
'db_name',
],
stdout: 'piped',
stderr: 'piped',
stdin: "piped",
})
await p.stdin.write(encoder.encode(await Deno.readTextFile('./dbDump.sql')))
await p.stdin.close()
await p.close()
But for some reason whenever I do it this way I get an error ERROR 1064 (42000) at line 145: You have an error in your SQL syntax; check the manual that corresponds to your MariaDB server version which does not happen when I perform the exact same command in the bash.
Could someone please explain me how it has to be done properly?
Without a sample input file, it's impossible to be certain of your exact issue.
Given the context though, I suspect that your input file is too large for a single proc.stdin.write() call. Try using the writeAll() function to make sure the full payload goes through:
import { writeAll } from "https://deno.land/std#0.119.0/streams/conversion.ts";
await writeAll(proc.stdin, await Deno.readFile(sqlFilePath));
To show what this fixes, here's a Deno program pipe-to-wc.ts which passes its input to the Linux 'word count' utility (in character-counting mode):
#!/usr/bin/env -S deno run --allow-read=/dev/stdin --allow-run=wc
const proc = await Deno.run({
cmd: ['wc', '-c'],
stdin: 'piped',
});
await proc.stdin.write(await Deno.readFile('/dev/stdin'));
proc.stdin.close();
await proc.status();
If we use this program with a small input, the count lines up:
# use the shebang to make the following commands easier
$ chmod +x pipe-to-wc.ts
$ dd if=/dev/zero bs=1024 count=1 | ./pipe-to-wc.ts
1+0 records in
1+0 records out
1024 bytes (1.0 kB, 1.0 KiB) copied, 0.000116906 s, 8.8 MB/s
1024
But as soon as the input is big, only 65k bytes are going through!
$ dd if=/dev/zero bs=1024 count=100 | ./pipe-to-wc.ts
100+0 records in
100+0 records out
102400 bytes (102 kB, 100 KiB) copied, 0.0424347 s, 2.4 MB/s
65536
To fix this issue, let's replace the write() call with writeAll():
#!/usr/bin/env -S deno run --allow-read=/dev/stdin --allow-run=wc
const proc = await Deno.run({
cmd: ['wc', '-c'],
stdin: 'piped',
});
import { writeAll } from "https://deno.land/std#0.119.0/streams/conversion.ts";
await writeAll(proc.stdin, await Deno.readFile('/dev/stdin'));
proc.stdin.close();
await proc.status();
Now all the bytes are getting passed through on big inputs :D
$ dd if=/dev/zero bs=1024 count=1000 | ./pipe-to-wc.ts
1000+0 records in
1000+0 records out
1024000 bytes (1.0 MB, 1000 KiB) copied, 0.0854184 s, 12.0 MB/s
1024000
Note that this will still fail on huge inputs once they exceed the amount of memory available to your program. The writeAll() solution should be fine up to 100 megabytes or so. After that point you'd probably want to switch to a streaming solution.
First, a couple of notes:
Deno currently doesn't offer a way to create a detached subprocess. (You didn't mention this, but it seems potentially relevant to your scenario given typical docker compose usage) See denoland/deno#5501.
Deno's subprocess API is currently being reworked. See denoland/deno#11016.
Second, here are links to the relevant docs:
docker-compose exec
CLI APIs > Deno.run
Manual > Creating a subprocess (Deno v1.17.0)
Now, here's a commented breakdown of how to create a subprocess (according to the current API) using your scenario as an example:
module.ts:
const dbUser = 'actual_database_username';
const dbPass = 'actual_database_password'
const dbName = 'actual_database_name';
const dockerExecProcCmd = ['mysql', '-u', dbUser, '-p', dbPass, dbName];
const serviceName = 'actual_compose_service_name';
// Build the run command
const cmd = ['docker', 'compose', 'exec', '-T', serviceName, ...dockerExecProcCmd];
/**
* Create the subprocess
*
* For now, leave `stderr` and `stdout` undefined so they'll print
* to your console while you are debugging. Later, you can pipe (capture) them
* and handle them in your program
*/
const p = Deno.run({
cmd,
stdin: 'piped',
// stderr: 'piped',
// stdout: 'piped',
});
/**
* If you use a relative path, this will be relative to `Deno.cwd`
* at the time the subprocess is created
*
* https://doc.deno.land/deno/stable/~/Deno.cwd
*/
const sqlFilePath = './dbDump.sql';
// Write contents of SQL script to stdin
await p.stdin.write(await Deno.readFile(sqlFilePath));
/**
* Close stdin
*
* I don't know how `mysql` handles `stdin`, but if it needs the EOT sent by
* closing and you don't need to write to `stdin` any more, then this is correct
*/
p.stdin.close();
// Wait for the process to finish (either OK or NOK)
const {code} = await p.status();
console.log({'docker-compose exit status code': code});
// Not strictly necessary, but better to be explicit
p.close();

Elephant io Undefined index : upgrades

Trying to reach ly node server with elephant io
But I fail with that error :
In Version1X.php line 194: Notice: Undefined index: upgrades
I added the dependency to my composer and the wissembly/elephant-io is well present in my vendor.
composer
"wisembly/elephant.io": "~3.0"
action
$client = new Client(new Version1X('http://localhost:1299/map'));
// open connection
$client->initialize();
$output->writeln("update des offres");
$offers = $this->em->getRepository('AppRefactoredBundle:app_Offer')->findAll();
foreach($offers as $offer){
if($offer->getUpdatedAt() < new DateTime('-2 hours')){
$offer->setCurrentStatus(2);
// send for server (listen) the any array
$client->emit('timeout', ['idUser' => $offer->getProfile()]);
}
}
Here is the doc: https://github.com/mairesweb/socket.io-elephant.io/blob/master/client.php
I don't require the autoload cause I can't find it, may be this is why...but I don't know how to do that.
Thanks for help.

When using Sqlite3 on Electron app , streams are undefined - Node's set-blocking module

We are working on a universal app, that must execute on mobile (ios, android), desktop (windows, mac, linux) adn in any browser.
For client database management, we want to use sqlite3 - for desktop- , in this case the app will be packaged using Electron (Atom shell) . The module bundler we are using is webpack, as the app is developed with Ionic 2 and Angular2.
We have installed sqlite3 well, as a project dependency, with node-pre-gyp generating the binary for the platform (in this case we are testing with Windows 7 64 bits)
We have a provider for Sqlite3, this is the code:
import * as sqlite3 from 'sqlite3';
import { IDatabaseProvider } from './database.provider';
export class Sqlite3DatabaseProvider implements IDatabaseProvider {
private _storage;
constructor() {
console.log('Initializing Sqlite3 Database Provider');
}
openDatabase() {
this._storage = new sqlite3.Database('v2App_sqlite3.db');
}
}
As you can notice, the line the creates a database is commented, as the app works OK like that. If I uncomment that line , we have this error:
Runtime Error
Cannot read property '_handle' of undefined
TypeError: Cannot read property '_handle' of undefined
at http://localhost:8100/build/main.js:210758:15
at Array.forEach (native)
at module.exports (http://localhost:8100/build/main.js:210757:36)
at Object.<anonymous> (http://localhost:8100/build/main.js:12580:1)
at Object.<anonymous> (http://localhost:8100/build/main.js:12873:30)
at __webpack_require__ (http://localhost:8100/build/main.js:20:30)
at Object.<anonymous> (http://localhost:8100/build/main.js:69457:11)
at Object.<anonymous> (http://localhost:8100/build/main.js:69638:30)
at __webpack_require__ (http://localhost:8100/build/main.js:20:30)
at Object.<anonymous> (http://localhost:8100/build/main.js:211548:72)
The curious of this, is that it's failing here, in the code of the set-blocking npm module:
module.exports = function (blocking) {
[process.stdout, process.stderr].forEach(function (stream) {
if (stream._handle && stream.isTTY && typeof stream._handle.setBlocking === 'function') {
stream._handle.setBlocking(blocking)
}
})
}
As stream is coming undefined ,when getting the _handle property gives the error.
But this code is only executed if I add the line that creates the sqlite3 database:
this._storage = new sqlite3.Database('v2App_sqlite3.db');
What is the relation between this module (set-blocking ) and sqlite3 ?, Why the streams are undefined when trying to create the sqlite3 database ?
Maybe as generates a file - the database - , using Node's outputStream, and still could not have been created that object - stream.
Any help?
Thanks in advance

Storm-R integration

I am trying to integrate my R script with Storm. The code for my Rbolt is:
public class RBolt extends ShellBolt implements IRichBolt {
public RBolt() {
super("Rscript", "storm_OR.R");
}
#Override
public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {
outputFieldsDeclarer.declare(new Fields("OR"));
}
#Override
public Map<String, Object> getComponentConfiguration() {
Config ret = new Config();
ret.setMaxTaskParallelism(1);
return ret;
}
}
I am getting the following error. Any help? I have made sure that the path variables have path of R and Rscript.
17469 [Thread-12-__system] INFO backtype.storm.daemon.executor - Preparing bolt __system:(-1)
17474 [Thread-12-__system] INFO backtype.storm.daemon.executor - Prepared bolt __system:(-1)
17480 [Thread-6] INFO backtype.storm.daemon.executor - Loading executor RBolt:[1 1]
17483 [Thread-6] INFO backtype.storm.daemon.executor - Loaded executor tasks RBolt:[1 1]
17491 [Thread-6] INFO backtype.storm.daemon.executor - Finished loading executor RBolt:[1 1]
17491 [Thread-6] INFO backtype.storm.daemon.worker - Launching receive-thread for 8d8a13de-5e87-4e14-b2c2-59b4dfc070c6:1027
17493 [Thread-14-RBolt] INFO backtype.storm.daemon.executor - Preparing bolt RBolt:(1)
17496 [Thread-15-worker-receiver-thread-0] INFO backtype.storm.messaging.loader - Starting receive-thread: [stormId: EventProcessing-1-1457335172, port: 1027, thread-id: 0 ]
17500 [Thread-14-RBolt] INFO backtype.storm.utils.ShellProcess - Storm multilang serializer: backtype.storm.multilang.JsonSerializer
17510 [Thread-14-RBolt] ERROR backtype.storm.util - Async loop died!
java.lang.RuntimeException: Error when launching multilang subprocess
at backtype.storm.utils.ShellProcess.launch(ShellProcess.java:64) ~[storm-core-0.9.2-incubating.jar:0.9.2-incubating]
at backtype.storm.task.ShellBolt.prepare(ShellBolt.java:99) ~[storm-core-0.9.2-incubating.jar:0.9.2-incubating]
at backtype.storm.daemon.executor$fn__5641$fn__5653.invoke(executor.clj:690) ~[storm-core-0.9.2-incubating.jar:0.9.2-incubating]
at backtype.storm.util$async_loop$fn__457.invoke(util.clj:429) ~[storm-core-0.9.2-incubating.jar:0.9.2-incubating]
at clojure.lang.AFn.run(AFn.java:24) [clojure-1.5.1.jar:na]
at java.lang.Thread.run(Thread.java:745) [na:1.7.0_67]
Caused by: java.io.IOException: Cannot run program "Rscript" (in directory "/tmp/933c85f3-f5b5-4a60-b342-7d4969b43d46/supervisor/stormdist/EventProcessing-1-1457335172/resources"): error=2, No such file or directory
This directory in tmp folder does not exist and is created on the fly. Any suggestions please.
UPDATE: Resolved this by creating another resources folder in the resources folder of the project such that the jar has a resources folder with the R script in it.
The whole purpose of "shell" components is to start as an independent process, therefore your script needs to implement multilang protocol.
Alternatively you can find a library that implements the protocol and has R integration, like FsStorm: it implements multilang and you can call R functions via R type provider.

Resources