Does teradata have an inbuilt way of generating ER diagrams of databases within it? If so, can you tell me how to do it?
I can't use third party tools.
I am not aware of any tools offered by Teradata that support the generation of ER diagrams. You will need to consider tools such as ERwin from Computer Associates or ER/Studio from Embarcadero. (Not endorsing either of these products, merely offering them as examples.)
There may be some other Open Source tools that support ER Diagrams but I am not familiar with them.
Related
tl;dr I want to deploy "live" model results in Python and R, and while Salesforce Einstein advertises this functionality for R and Python, I have only found support for Python. Shiny is too expensive to justify for our limited R-language requirements. Does Einstein R support actually exist?
UPDATE: Tableau has a separate solution from Einstein Analytics that hosts both R and Python - see answer below. Not a feature-rich direct competitor to Shiny, but that's not our use-case.
According to the documentation for Salesforce Einstein Analytics Plus (aka Tableau CRM AI Analytics), data scientists can upload (operationalize) their Python, R, and Matlab code, as described here:
https://www.tableau.com/solutions/ai-analytics (see the section on "Data Science" at the bottom of the page).
I signed up for a trial of Einstein Analytics Plus, and found a link to the "Model Manager." Using Model Manager to deploy Python-language models is well-documented here:
https://help.salesforce.com/s/articleView?id=sf.bi_edd_model_upload_prepare.htm&type=5
For Python, this seems to match the advertised functionality. But there is no indication of how to deploy R language models, which may be part of my team's use case.
I would like to find the equivalent method for deploying an R-language model in Einstein. Particularly, is there some other Salesforce / Tableau product I should try, or is this a feature that is simply not available in the trial version. Unlike Python deployment, searching the documentation has not yielded answers.
Alternatively, we're only interested in Einstein R support is because it appears to be about 1/10 the cost of Shiny, which is hideously expensive. So any recommendations regarding lightweight alternatives to Shiny would also be helpful.
TIA for anyone who can shine a light on this problem.
ANSWER: There is actually a separate feature in Tableau that is different from Einstein Analytics which supports both R and Python, documentation here:
https://help.tableau.com/current/prep/en-us/prep_scripts.htm
I find that there's very little documentation on how to extract SAP tables into R.
I'm not talking about SAP HANA.
Currently, it's very troublesome that I need to manually extract SAP tables using a GUI interface, export them into tabular format. Then only I can import them using my R script.
The current solution I'm exploring is to have my SAP colleagues to export those SAP tables into SQL database, then I can query the tables from R.
Ideally I want to cut this seemingly unnecessary step of having the SAP tables exported into a database.
For SAP R/3 systems (or what you call ECC), your best bet would be executing remote function calls (i.e. RFC).
Normally these would be supported by open source interfaces for at least the more recent versions (e.g. 4.6 or above).
However, they are fairly scarce and I know only of one such implementation in R - this is the RSAP. You'd also need to download NW RFC SDK, and there may be further requirements based on your OS (e.g. what Visual C++ you'd need for Windows, etc.).
There's also a slightly more widely recognised equivalent in Python, the PyRFC.
On the other hand, you may try Robotic Process Automation (RPA) to interact with GUI in an automated way. One of the options is UiPath but there are others. This way you could configure the automation of table extraction - at the same time you can also call R scripts directly from the RPA.
Overall - to be honest - the solution with extracting tables into a separate database does seem to be the best alternative (compared to what I've described above).
Note: The above presumes that - for any reason, usually security - you cannot access the database underlying ECC directly through ODBC calls - otherwise the instructions for connecting and calling SQL from R are the same as for HANA or similar.
Consider using RODBC. This package allows adding different ODBC sources and use them in R Studio.
Follow this article and don't bug to word "HANA", this approach allows using any database, not only HANA.
do you know any good software that manages multigraph? My aim is to display it, compute some metrics on it (when well defined), import/export from common network formats (GraphML...) and so on.
Something like Gephi or NetSocV but with multigraph support.
Linux support would be great.
I found NetworkX for Python and it seems good, but I was wondering if there are alternatives available.
Any recommendations?
Very often I get into the projects that have requirements of transferring file data into table. And almost always I've worked at ODI (Oracle Data Integrator) only.
I want to know what are the different ETL tools available and how are they different from ODI and what are the restrictions in each case (like file size limit or column size restriction or processing time etc).
I wish somebody could help.
If somebody can share personal experience on these tools, that would be welcome too. Thanks!
I'm working on the same type of projects that you're in.
Right now I'm working with IBM DataStage. It seems like a good and powerful tool, but it's lacking a good documentation and a strong community.
There's also Pentaho, I have no experience about it, but it seems pretty popular and it's also open source
Has anyone successfully used Rhapsody in a DO-178 avionics environment? That is, working with the FAA/DER process to provide artifacts to them and have them approved. Since it is my understanding that Rhapsody isn't a certifiable MDD tool, I was curious if there were other mitigating factors.
If you were successful so, what steps did you take in order to be able to accomplish this?
Thanks for any feedback and insights.
I have used Rhapsody on a project that was developed in accordance with (but not certified) to DO-178B level D. The requirements were managed in DOORS and linked into Rhapsody using the Rhapsody Gateway tool, which worked reasonably well. This was important as traceability is a key part of 178B.
The software was modelled in Rhapsody and the code then generated manually. Manual code generation was chosen as auto-generation of code would then require Rhapsody to be qualified as a development tool to comply with 178B. I don't know if IBM provide any 178B certification for Rhapsody.
Verification of the software against requirements was performed using a bespoke test tool, and for this we had to perform some significant testing of the tool in order to qualify it as a verification tool.
Your question is quite hard to answer as you don't include any information on what level of 178B you are working to, what tools you are using/planning to use (other than Rhapsody), or whether you are intending to auto generate code, etc.
Hope this is of some help.
I have experience using Rhapsody C++ for DO-178B Level A/B compliant project.
Auto generated code is verified in accordance with the coverage requirements, including MC/DC coverage, for the proper level. Since the generated code are fully verified with rigorous static/dynamic tests and manual reviews, as if they were hand coded, the Rhapsody tool qualification was not mandatory.
We have put much effort in customizing Rhapsody code generation properties to generate only the needed code such as ctor/dtors and get/setters, and to avoid library functions which are not deterministic or the ones with dynamic memory allocations.
We were able to fully utilize round-trip engineering so that the Rhapsody model files, not the code, are version-controlled since the model contains all the code.
Rhapsody UML should be considered for developing reusable and portable software architecture.
Rhapsody is being used in our Level A/C/D project with Arinc 653. Since Output of Rhapsody(Auto code generators) are being verified.
Hence, Qualifying Rhapsody is not necessary. Rhapsody gives advantages in Traceability and generation or modifying Test Scripts by updating just "Tags" field.
So the entire Test script or traces in Test script need not to be modified.