Tableau Linux server, does it not recognize the user ODBC settings - odbc

Connecting data source via ODBC driver to the tableau server seem to not work at a user level.
I have odbc.ini and obdcinst.ini files on the user home directory. $ODBCINI and $ODBCINSTINI Environment variables are set pointing to respective files.Tableau still fail to pick the data source.
Workbook published from the desktop is rendered only on placing the odbc.ini and odbcinst.ini into /etc/ or tableau user directory (/var/opt/tableau/tableau_server/).
Although it fixes the issue, there are multiple users on the system and we do not want to configure at a system level affecting all the users.
https://community.tableau.com/docs/DOC-17674 suggests to configure ODBCSYSINI in /etc/systemd/system/tabsvc_0.service. The latest version of tableau (2018.3) seem to have the conf file at /var/opt/tableau/tableau_server/.local/share/systemd/user/tabsvc_0.service. How can we configure it to use the settings for each user and restart this service?

Related

RODBC through front end to SharePoint

I have an R script that uses the RODBC package to connect to an .accdb file allowing me to pull data directly into R. The .accdb file was recently reconfigured so the tables reside on SharePoint and we access the table through the front-end .accdb file. Now, when I connect to the front-end I am no longer able to pull data from the tables/queries. If I run sqlFlecth() I get an error "42S02 - 1305" that says it could not find the object InsertTableOrQueryName and says if the object is not a local object to check my network connections or contact the server administrators. I have a network connection - I can open the front-end and run the queries/access the tables. My administrators aren't any help. Any know how I can get this working again so I don't have to keep opening the front-end running the queries and saving to an intermediate Excel file or something?

Can RDS (Mysql in my case) endpoint be connected with 2 instances in different availability zones?

What am I trying to achieve?
1.Trying to build a fault-tolerant WordPress website.
2.Tried installing the webserver on one AZ with Muti-AZ RDS deployment.It was quite successful.
Set up is as follows
AZ-1 Public subnet - Launched one ec-2 instance, Installed httpd, PHP, PHP-MySQL, WORDPRESS.
AZ-1 Private subnet - Launched a Multi-AZ RDS instance
Problem Encountered:
When I Wanted to expand to another availability zone for fault tolerance.
Launched another ec-2 instance in different availability zone [AZ-2] and installed httpd, PHP, PHP-MySQL, WORDPRESS
I DID NOT launch an another RDS.I wanted to connect to the RDS in [AZ-1]coz its already a Multi-AZ, So wanted to have the fault tolerance set up only for the Web server. I was able to install WordPress on AZ-2 public subnet, but I was unable to connect to the
RDS[MYSQL]endpoint in AZ-1.
Getting the error message.
"Already installed.You appear to have already installed WordPress. To reinstall please clear your old database tables first".
"Already installed.You appear to have already installed WordPress. To
reinstall please clear your old database tables first".
This means your second web server can successfully connect to the RDS instance. Instead of trying to "install" WordPress, just copy all your WordPress files from the first web server and you'll be fine.

DeployR Open - Restrict file system access

According to the DeployR documentation each session is locked down:
"By default, R sessions executing on the DeployR grid are not authorized to access files or directories outside of the R working directory"
However, I have been able to upload a simple script using a user that has Basic_User role and that reads a file from the c:\ drive on windows.
t <- read.table("C:\\MyFile.csv")
t
How do you lock down access to the file system? I seem to have missed something.

Samba Share directory outside user directory

I set up a machine with Ubuntu 14.4.1 yesterday and I am trying to use it as remote storage for a project I am working on remotely with a few people. I know nothing about hosting servers, so I am attempting to avoid the issue entirely by just treating it like a local area network using Hamachi.
The Ubuntu machine has 2 hard drives - a boot drive on which Linux is installed and a larger data drive. I am attempting to share a directory on the data drive via samba so that it can be accessed via Hamachi by windows 7 machines.
I am able to see the directories that I have shared, but when I try to enter them, I get a permission denied error. When I share a directory within my /home/user/ directory, it works fine. Is there any way that I can share a directory on my data drive?
Perhaps I could make a symbolic link from my user directory to the data drive? Would that actually work? I am not familiar.
If it is NTFS, you can't change the permissions, which will be the issue.

Best way to install web applications (e.g. Jira) on Unixes?

Can you throw some points on how it is a best way, best practice
to install web application on Unixes?
Like:
where to place app and its bases and so for,
how to configure to be secure and easy to backup,
etc
For example I know such suggestion -- to set uniq user for each app.
App in question is Jira on FreeBSD, but more general suggestions are also welcomed.
Here's what I did for my JIRA install on Fedora Linux:
Create a separate user to run JIRA
Install JIRA under the JIRA user's home directory
Made a soft link "/home/jira/jira" pointing to the JIRA installation directory (the directory as installed contains the version number, something like /home/jira/atlassian-jira-enterprise-4.0-standalone)
Created an /etc/init.d script to run JIRA as a service, and added it to chkconfig so that it runs at system startup - see these instructions
Created a MySQL database for JIRA on a separate data volume
Set up scheduled XML backups via the JIRA admin interface
Set up a remote backup script to dump the MySQL database and copy the DB dump and XML backups to a separate backup server
In order to avoid having to open extra firewall ports, set up an Apache virtual host "jira.myhost.com" and used mod_proxy to forward requests to the JIRA URL.
I set everything up on a virtual machine (an Amazon EC2 instance in my case) and cloned the machine image so that I can easily restart a new instance if the current one goes down.

Resources