Installing apex in solaris - oracle11g

I neva worked with solaris or Linux before, and would like to find out how i can install oracle application express using command line in solaris, after i have installed my oracle database 11g in solaris aswell.
I already have an idea of how to install the database.

i agree with the comments above, the apex installation guide on the oracle site is easy to follow, basically, you will have to:
execute scripts in the oracle database (first create some tablespaces, then run an sql script that will install the apex module)
deploy a file containing an apex "listener" that you will have to quickly configure, and run a WAR (java) file to run apex
and that's it :)

Installation of Application Express (APEX) is largely Operating System agnostic. The process is based around running a number of scripts. When performing APEX installations, my primary tool is SQL Plus.
Depending on the version\edition of the Oracle Database you install, you probably already have a version of APEX ready to use. You can check which version of APEX your database has by running the following sql statement:
select version_no from apex_release
More information about installing Apex can be found in the documentation.

the documentation on the oracle website is well done.
i installed apex on Solaris 10 and it works fine
basically, you just need to execute a few SQL scripts (create some users, import data), and afterwards run a java program to start the apex listener
check the documentation, it should be understandable even without too much of Solaris knowledge

Related

An up-to-date SQLite ODBC driver

Since one of my SQLite db file has been managed by the latest SQLite version (3.39.2), I am no longer able to access the tables through the ODBC driver. There is no error when connecting to the database, but the list of tables remains empty.
I have a suspicion that this is because it is seriously behind, since it's built on version 3.32.3.
Is there any other more recent ODBC driver for SQLite available out there?
I have played with the SQLiteODBC driver last year (I used it with Excel 2002 and 2016) and patched the official scripts to enable builds that include current SQLite release. The details are here.
I have just recompiled both x32 and x64 SQLiteODBC installers for Windows and uploaded them to my repo. You can install those or, if you wish, follow my docs and build the driver yourself (you would still need to use several patched scripts included in my repo).
Besides including the current SQLite release, I also enabled most of the extra extensions, which are not included in the official ODBC builds.

Import and explore Oracle .dmp file

I have a .dmp file (export from an Oracle DB) I need to get data from. I have installed Oracle 11g Express Edition successfully, with user SYSTEM and imp command I imported the file, with some warnings though, but now after hours of searching the Internet, I still have no idea how to explore the DB created by that file.
Can explain to me how to do that or at least point me to some relevant documentation, please? This is the first time I use Oracle.
Thanks!
Update: I already tried with Oracle SQL Developer. I use the user SYSTEM that was created automatically when I installed XE

Teradata sample database

I have installed vmware player and added the disk file to it. Teradata studio is already installed in it. Now, I want create a database of good size to perform bulk load types in Teradata.
Where can I get the sample database?
I assume that you are asking about the software part to create an instance of Teradata database. If you have VMware downloaded from Teradata for evaluation purpose, it is already installed and configured. You just have to connect with your Studio.
If you are trying to create a virtual environment on your own, then you need to download the Teradata software and install in machine which will create a database. After installing the software you can configure and connect with Studio.
Teradata Express comes with toy data (Samples database) but only up to version 14. From version 15 it is (oddly) removed.
Your VM-Ware is fully configured. Sart your TD-Studio and connect to it.
From VM-Ware's TD-Studio it is TDPID: localhost od 127.0.0.1 User dbc pass dbc
There are no more sample databases in Teradata Express, but I published several data sets on GitHub, see the blog on Teradata Community

Presto Interpreter in Zeppelin on EMR

Is it possible to add Presto interpreter to Zeppelin on AWS EMR 4.3 and if so, could someone please post the instructions? I have Presto-Sandbox and Zeppelin-Sandbox running on EMR.
There's no official Presto interpreter for Zeppelin, and the conclusion of the Jira ticket raised is that it's not necessary because you can just use the jdbc interpreter
https://issues.apache.org/jira/browse/ZEPPELIN-27
I'm running a later EMR with presto & zeppelin, and the default set of interpreters doesn't include jdbc, but it can be installed using a ssh to the master node and running
sudo /usr/lib/zeppelin/bin/install-interpreter.sh --name jdbc
Even better is to use that as a bootstrap script.
Then you can add a new interpreter in Zeppelin.
Click the login-name drop down in the top right of Zeppelin
Click Interpreter
Click +Create
Give it a name like presto, meaning you need to use %presto as a directive on the first line of a paragraph in zeppelin, or set it as the default interpreter.
The settings you need here are:
default.driver com.facebook.presto.jdbc.PrestoDriver
default.url jdbc:presto://<YOUR EMR CLUSTER MASTER DNS>:8889
default.user hadoop
Note there's no password provided because the EMR environment should be using IAM roles, and ppk keys etc for authentication.
You will also need a Dependency for the presto JDBC driver jar. There's multiple ways to add dependencies in Zeppelin, but one easy way is via a maven groupid:artifactid:version reference in the interpreter settings under Dependencies
e.g.
under artifact
com.facebook.presto:presto-jdbc:0.170
Note the version 0.170 corresponds to the version of Presto currently deployed on EMR, which will change in the future. You can see in the AWS EMR settings which version is being deployed to your cluster.
You can also get Zeppelin to connect directly to a catalog, or a catalog & schema by appending them to the default.url setting
As per the Presto docs for the JDBC driver
https://prestodb.io/docs/current/installation/jdbc.html
e.g. As an example, using Presto with a hive metastore with a database called datakeep
jdbc:presto://<YOUR EMR CLUSTER MASTER DNS>:8889/hive
OR
jdbc:presto://<YOUR EMR CLUSTER MASTER DNS>:8889/hive/datakeep
UPDATE Feb 2018
EMR 5.11.1 is using presto 0.187 and there is an issue in the way Zeppelin interpreter provides properties to the Presto Driver, causing an error something like Unrecognized connection property 'url'
Currently the only solutions appear to be using an older version in the artifact, or manually uploading a patched presto driver
See https://github.com/prestodb/presto/issues/9254 and https://issues.apache.org/jira/browse/ZEPPELIN-2891
In my case using an old reference to a driver (apparently must be older than 0.180) e.g. com.facebook.presto:presto-jdbc:0.179 did not work, and zeppelin gave me an error about can't download dependencies. Funny error but probably something to do with Zeppelin's local maven repo not containing this, not sure I gave up on that.
I can confirm that patching the driver works.
(Assuming you have java & maven installed)
Clone the presto github repo
Checkout the release tag e.g. git checkout 0.187
Make the edits as per that patch https://groups.google.com/group/presto-users/attach/1231343dbdd09/presto-jdbc.diff?part=0.1&authuser=0
Build the jar using mvn clean package
Copy the jar to the zeppelin machine somewhere zeppelin user has permission to read.
In the interpreter, under the Dependencies - Artifacts section, instead of a maven reference use the absolute path to that jar file.
There appears to be an issue passing the user to the presto driver, so just add it to the "default.url" jdbc connection string as a url parameter, e.g.
jdbc:presto://<YOUR EMR CLUSTER MASTER DNS>:8889?user=hadoop
Up and running. Meanwhile, it might be worth considering Athena as an alternative to Presto give it's serverless & is effectively just a fork of Presto. It does have limitation to External hive tables only, and they must be created in Athena's own catalog (or now in AWS Glue catalog, also restricted to External tables).
Chris Kang has a good post on doing that in spark-shell, http://theckang.com/2016/spark-with-presto/. I don't see you wouldn't be able to do that in Zeppelin. Another helpful post is making sure you have the right Java version in EMR, http://queirozf.com/entries/update-java-to-jdk-8-on-amazon-elastic-mapreduce. The current Presto version as of writing only runs on Java 8. I hope it sets you in the right direction.

use sqlite with node.js - working module

I'm working on a node.js project and I would like to use some kind of database to store application data. I've searched for various DBMS and I've selected SQLite because I want my project to be able to run without installing heavy DBMS and because I want it to run both on linux and windows. however, all modules that we found, that connect node with SQLite are either very old, or cannot be installed due to bugs.
so, I want to ask if nowadays, 2012, there is a completely functional node.js module for SQLite.
I did a project last July (2012) in Node.JS v0.8.8, using the node-sqlite module (see https://github.com/developmentseed/node-sqlite3) and it worked just fine.
And, I successfully deployed and used this module on Heroku. However, the deployed portion of the project only involved reading out of the SQLite database – not writing. I suspect that writing into a SQLite deployed on Heroku would cause some issues, because it involves modifying the SQLite file in the file system, which AFAIK Heroku doesn't let you do.

Resources