installing Realm through Cocoa pods fail - realm

i m getting this error below while installing Realm through Cocoapods . What's the problem here? I rarely get answers on stack overflow although everybody advices here but I hope this time I can get an answer !
Terminal Screenshot

Related

Set up rstudio-server on macOS

I want to set up rstudio-server on an iMac with support for multiple users and remote login. I followed the steps in the INSTALL tutorial: I built the source, set up the configuration files and the launchd daemon. At first, it works fine, but after some time, I get these warnings/errors when I plot:
2022-06-09 08:02:29.438 rsession[3050:139329] XType: failed to connect - Error Domain=NSCocoaErrorDomain Code=4099 "The connection to service named com.apple.fonts was invalidated: failed at lookup with error 3 - No such process." UserInfo={NSDebugDescription=The connection to service named com.apple.fonts was invalidated: failed at lookup with error 3 - No such process.}
2022-06-09 08:02:29.438 rsession[3050:139329] Font server protocol version mismatch (expected:5 got:0), falling back to local fonts
2022-06-09 08:02:29.438 rsession[3050:139329] XType: unable to make a connection to the font daemon!
2022-06-09 08:02:29.438 rsession[3050:139329] XType: XTFontStaticRegistry is enabled as fontd is not available.
Then I can't plot any more unless I restart R and re-run my code. Do you know what could be the issue? I could not get any help when opening an issue on the rstudio-server github since MacOS is not officially supported.
I was also looking at running rstudio-server via docker, but I couldn't find a good way to map the user namespace from macOS to the container.
Any help or suggestion would be greatly appreciated!
EDIT: It seems I was able to solve the issue by launching the fontd daemon with:
sudo launchctl load -w /System/Library/LaunchAgents/com.apple.fontd.useragent.plist
This seems like an issue with the MacOS font daemon, not with RStudio itself.
Someone reported a similar issue on PhantomJS. Rebooting resolved it for them.
This answer reported the same error for a different build, and they were able to resolve it by installing the correct "Apple Worldwide Developer Relations Certification Authority" in Keychain:
The one I had had an expiration date of February 2023. I deleted that one and went here, downloaded the one called "Worldwide Developer Relations - G3 (Expiring 02/20/2030 00:00:00 UTC)", then retried the build and it worked.

Telepresence Connection Error - Traffic Manager version unsupported, must be 2.4.5 or higher while it is 2.6.5

I have started to face this problem. While trying to connect, I am facing an error stating that my traffic-manager version is 2.1.5 and it should be at least 2.4.5.
"telepresence connect" command checks for new versions and modifies it if there is any new version exists. So I am thinking that started to create a problem. Because I was using it as normal.
When I check the connector.log file these two lines create the problem.
connector/session : Existing Traffic Manager 2.6.5 not owned by cli or does not need upgrade, will not modify
connector/session : failed to connect to root daemon: rpc error: code = Unknown desc = unsupported traffic-manager version 2.1.5. Minimum supported version is 2.4.5
So somehow I have two versions now while checking for the update it hits 2.6.5 but while trying to run it tries with 2.1.5. Trying to uninstall telepresence but it also faces the same problem and I couldn't locate and delete traffic-manager 2.1.5. My OS is Windows 11.
Because of that, I am kind of blocked with my tests. Any help will be well appreciated. Thanks!
After asking the question, a new version arrived, if anyone encountered this problem, please update telepresence to 2.6.6. It is fixed now.

Issues installing Plotly Dash for R

I'm trying to install the newly released Plotly Dash package for R as described on plotly website https://dashr.plot.ly/installation, i.e. running
library(devtools) # devtools: Tools to Make Developing R Packages Easier
# The following statement will also install dashCoreComponents,
# dashHtmlComponents, and dashTable
install_github('plotly/dashR') # The core dash backend
This is what it returns, 'unknown package' any help much appreciated
Error: Failed to install 'unknown package' from GitHub:
HTTP error 403.
API rate limit exceeded for 86.134.113.102. (But here's the good news:
Authenticated requests get a higher rate limit. Check out the
documentation for more details.)
Just installed the package to reproduce the error, unfortunately can't be able to reproduce it. However, a potential solution along "error 403 and
API rate limit exceeded ..." could be to add your github authorization token in your .Rprofile and that is what the "But here's the good news:" message inside the bracket is trying to inform you. Have a look at the actual project github site Dash for R if you haven't already. That might as well provide you an input to solving the issue.

multiple version of spark on CDH5.10 is failing to launch spark-submit

I have installed spark with 2.0 on CDH5.10 By following the link https://www.cloudera.com/documentation/spark2/latest/topics/spark2_installing.html
after all configuration when I hit spark2-submit --version it gives me correct version which is 2.0
however when I submit a spark job . First it says
Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/hadoop/fs/FSDataInputStream
This is clearly indicating that hadoop libs are not in classpath. My question is it something wrong with my installation of spark 2. ? also once we add jars with sparkExtralibCLasspath for driver and core then it says SPARK_HADOOP_CONF Is not set.
How can I verify my installation is correct ?
I am also trying to understand where are my spark2 conf dirs
I saw few previous question on stackoverflow like https://community.cloudera.com/t5/Cloudera-Manager-Installation/CHD-5-7-spark-shell-java-lang-ClassNotFoundException-org-apache/td-p/42209 and NoClassDefFoundError com.apache.hadoop.fs.FSDataInputStream when execute spark-shell but this doesnt help
I am using spark2-shell and spark2-submit command
some more investigation with https://community.cloudera.com/t5/Cloudera-Manager-Installation/CDH-5-5-pyspark-java-lang-NoClassDefFoundError-org-apache-hadoop/td-p/34424 shows might be If I can correctly set SPARK_EXTRA_LIB_PATH for spark2 then I can fix this issue. can somebody guide me please. Thanks

Torch / cunn install error - file already has LC_RPATH

I'm trying to install cunn for Torch but make install fails. Cuda is 7.5.27 because 8 doesn't support OSX 10.10 anymore.
I've been trying to debug this for a while, but I'm running out of things to try.
While searching solution I saw few others having a similar problem (https://github.com/TRIQS/cthyb/issues/8, https://github.com/jcjohnson/neural-style/issues/134) but didn't find any solution that would fix this. It seems that this is a general problem related to OSX/Xcode and not necessarily cunn.
All help is much appreciated.
$ luarocks install cunn
...
...
-- Installing: /Users/foo/torch/install/lib/luarocks/rocks/cunn/scm-1/lib/libTHCUNN.so
error: /Applications/Xcode.app/Contents/Developer/Toolchains/XcodeDefault.xctoolchain/usr/bin/install_name_tool: for:
/Users/foo/torch/install/lib/luarocks/rocks/cunn/scm-1/lib/libTHCUNN.so (for architecture x86_64) option "-add_rpath /usr/local/cuda/lib"
would duplicate path, file already has LC_RPATH for: /usr/local/cuda/lib
The issue went away when I did ./clean.sh in ~/torch followed by ./update.sh. I still don't have any idea what was the problem.

Resources