Hi I am trying to connect to Datastore using the below command
DatastoreOptions options = DatastoreOptions.builder().projectId("project-id").authCredentials(AuthCredentials.createForJson(new FileInputStream("c:\GoogleCredentials\.json"))).build();
I have created a folder GoogleCredentials in my C: Drive and downloaded the Service account's JSON keyfile into this folder
But I am getting the below error:
Line 38, Column 67: No applicable constructor/method found for actual parameters "java.lang.String"; candidates are: "com.google.cloud.ServiceOptions$Builder com.google.cloud.ServiceOptions$Builder.projectId(java.lang.String)"
Related
I try to connect to Azure Data Lake using Airflow. I use Airflow connection via the Web UI.
When I try to connect using the test button, I get an error Bad Request. As seen below
I use the correct UUIDs. These UUIDs have been verified in other cases. I also checked the firewall.
When I execute the DAG, I use the Azure Data Lake connection id to check if a file exists: If I apply the method as described here: What is the best way to check if a file exists on an Azure Datalake using Apache Airflow?
This is the error I get
[2022-05-06, 17:27:33 UTC] {log.py:127} ERROR - 99ec1d77-e91c-4fd3-a1c7-fa751ca1e779 - OAuth2Client:The token response from the server is unparseable as JSON: ***
Traceback (most recent call last):
File "/opt/airflow/lib/python3.8/site-packages/adal/oauth2_client.py", line 168, in _validate_token_response
wire_response = json.loads(body)
File "/usr/lib/python3.8/json/init.py", line 357, in loads
return _default_decoder.decode(s)
File "/usr/lib/python3.8/json/decoder.py", line 337, in decode
obj, end = self.raw_decode(s, idx=_w(s, 0).end())
File "/usr/lib/python3.8/json/decoder.py", line 355, in raw_decode
raise JSONDecodeError("Expecting value", s, err.value) from None
json.decoder.JSONDecodeError: Expecting value: line 3 column 1 (char 4)
[2022-05-06, 17:27:33 UTC] {log.py:127} ERROR - 99ec1d77-e91c-4fd3-a1c7-fa751ca1e779 - OAuth2Client:Error validating get token response: ***
Traceback (most recent call last):
File "/opt/airflow/lib/python3.8/site-packages/adal/oauth2_client.py", line 238, in _handle_get_token_response
return self._validate_token_response(body)
File "/opt/airflow/lib/python3.8/site-packages/adal/oauth2_client.py", line 168, in _validate_token_response
Authenticating to Azure Data Lake is by token credentials i.e. add specific credentials (client_id, secret, tenant) and account name to the Airflow connection.
Information about how to set it up can be found in this doc.
You can see code example in the source code test function.
Other method of authentication are currently not supported.
I was trying to get the connection running using the Airflow implementation. My impression was that it was buggy and did not work out well. The above situation happened with Airflow 2.2.5. When I upgraded to Airflow 2.3.0, the test button was grayed out.
The final solution was to use Access Tokens instead.
I have been trying to find help for this problem and there hasn't been too much success. I keep getting this error. I was following this guide: https://cloud.google.com/community/tutorials/cloud-iot-gateways-rpi and I haven't been able to get past step 14.
source run-gateway
Creating JWT using RS256 from private key file rsa_private.pem
on_publish, userdata None, mid 1
Unable to find key 1
connect status False
on_connect Connection Refused: not authorised.
on_disconnect 5: The connection was refused.
connect status False
connect status False
connect status False
^CTraceback (most recent call last):
File "./cloudiot_mqtt_gateway.py", line 356, in <module>
main()
File "./cloudiot_mqtt_gateway.py", line 284, in main
time.sleep(1)
I was able to successfully get the gateway running but had to manually modify the script. Make sure that you have updated the run-gateway shell script to point to your registry ID and project ID.
If any of the parameters are incorrect (e.g. device, project, region) then your device will be disconnected from the device bridge.
I am a new user of OrientDB. I tried to use OrientDB Console to create a new database. The steps I followed are from OrientDB official site: http://orientdb.com/docs/last/Console-Command-Create-Database.html. But I got errors.
1) run OrientDB server on the local PC: bin/server.bat
2) run console: bin/console.bat
3) connect to remote server:
orientdb> CONNECT remote:localhost root root
4) create a new database and then comes the error:
orientdb {server=remote:localhost/}> CREATE DATABASE REMOTE:localhost/myremotetestdb root root PLOCAL
Creating database [REMOTE:localhost/myremotetestdb] using the storage type
[plocal]...
Error: java.lang.IllegalArgumentException: Storage type 'plocal' is
different by storage type in URL
orientdb {server=remote:localhost/}>
I am trying to install Extensible Service Proxy on my compute instance. I am following this guide for installing the ESP Nginx service: https://cloud.google.com/endpoints/docs/quickstart-compute-engine#running_the_extensible_service_proxy. I was able to install the ESP service without problems but when I try to start the service with the command service nginx start, the service does not start up.
First it gave this error in /var/log/nginx/error.log:
Traceback (most recent call last):
File "/usr/sbin/start_esp.py", line 48, in <module>
from mako.template import Template
ImportError: No module named mako.template
The error went away after I installed the mako template module using the command pip install mako.
Now it is giving this error:
INFO:Fetching the service name from the metadata service
ERROR:Fetching service name failed (status code 404)
Any help would be much appreciated. Thanks
Did you forget to put in your service name in your metadata?
From https://cloud.google.com/endpoints/docs/quickstart-compute-engine:
In the Metadata section and add the following Endpoints metadata key/value pairs:
Specify endpoints-service-name as a key and YOUR-PROJECT-ID.appspot.com as its value, replacing YOUR-PROJECT-ID with your project ID.
Click Add item.
Specify endpoints-service-version as a key and the service version returned when you deployed as the key's value.
I am trying to install the latest version of swift following instructions from http://docs.openstack.org/icehouse/install.../general-installation-steps-swift.html. I am able to authenticate with keystone and also able to successfully run the command swift stat. But, when I run the command swift upload myfiles temp, I get the following error
Error trying to create container 'myfiles': 404 Not Found: {"error": {"message": "The
resource could not be found.", "c
Object PUT failed: 9.109.124.109:5000:5000/v2.0/myfiles/temp 400 Bad Request
[first 60 chars of response] {"error": {"message": "Expecting to find application/json
in
In /var/log/syslog, I find the following information:
May 28 18:11:40 datafed3 account-server: ERROR __call__ error with PUT /sdb1/100869
/AUTH_system/myfiles : #012Traceback (most recent call last):#012 File "/usr/lib
/python2.7/dist-packages/swift/account/server.py", line 284, in __call__#012 res =
method(req)#012 File "/usr/lib/python2.7/dist-packages/swift/common/utils.py", line
2217, in wrapped#012 return func(*a, **kw)#012 File "/usr/lib/python2.7/dist-
packages/swift/common/utils.py", line 837, in _timing_stats#012 resp = func(ctrl,
*args, **kwargs)#012 File "/usr/lib/python2.7/dist-packages/swift/account/server.py",
line 128, in PUT#012 req.headers['x-bytes-used'])#012 File "/usr/lib/python2.7/dist-
packages/swift/account/backend.py", line 210, in put_container#012 raise
DatabaseConnectionError(self.db_file, "DB doesn't exist")#012DatabaseConnectionError:
DB connection error (/srv/node/sdb1/accounts/100869/80d/62816079be0fc97a4557f52b3b12380d
/62816079be0fc97a4557f52b3b12380d.db, 0):#012DB doesn't exist
One situation may cause this problem is: when create tenant, one or more storage node is down. then when you upload an object, proxy get 404 from at least one storage node.
On my test, even the storage node are all up after tenant creation, 404 error still exist. So, make sure all storage nodes are up, and create another tenant to test.