according to here: http://watwp.codeplex.com/wikipage?title=Architecture%20Diagrams
The SQL Azure OData Service is a sample WCF Data Service built on top
of a SQL Azure (or SQL Server) database using Entity Framework 4.1
Code First.
The current version of this service only supports Read operations and,
in addition to exposing the SQL Azure database as an OData feed, it
adds a security layer to manage authentication / authorization.
so does it mean the my windows phone app will only be able to read from sql azure and not write to ? or can i do it by creating a data service on the asp.net server ?
i'm a little confused.
What this is saying is that the sample OData service that they provided only implements read operations. If you want read and write, you're going to have to roll that yourself.
Related
Following the advice in this article, I have configured my applications to use region-specific Gremlin endpoints so that reads and writes are always directed to the master replica in the same data centre (the Cosmos DB account is multi-master and the applications are deployed to every region on AKS). My question is this: in the event of a regional Cosmos DB outage, what will the behaviour be when using region-specific Gremlin connection strings? Will applications that reference a regional endpoint that is affected by an outage be automatically redirected to a region where the Cosmos replica is healthy?
This depends upon the client SDK the application is using to connect and the connection string logic. If the application connection string is pointed to the .NET SDK URI, then you will want to implement either the .NET SDK v2 or .NET SDK v3 multi-master functionality. If you are using the Gremlin Endpoint, please follow the specific guidance: Regional endpoints for Cosmos DB for Graph Accounts
Once that is configured correctly, in the event of an outage, the routing will automatically be redirected to an available write region.
I have a .NET Core 3 Web API that has some external dependencies like Azure SignalR and Azure CosmosDb. I have create a Monitoring Controller which I want to check if the service is healthy by for example pinging those two dependencies.
I haven't found any documentation about how to do this. I know that for example, for Table Storage I can call a method "If table exists" so if it returns true, I know the service can communicate with that dependency.
There is a way to call those dependencies?
PS: I don't want to use 3rd party packages to perform this, I want to know if those services provide this kind of methods.
I know that for example, for Table Storage I can call a method "If table exists" so if it returns true, I know the service can communicate with that dependency.
There is a way to call those dependencies?
If you'd like to achieve same for Azure SignalR and Azure CosmosDb, you can try:
1) To check Azure SignalR service, can use management API to get details of Azure SignalR service.
2) To check Azure CosmosDb, can use this API to check if the database you want is existing.
I created Azure CosmosDb database and container for my documents.
I use NiFi as a main data ingestion tool and want to feed my container with documents from NiFi flow files.
Can anybody please share a way to post flowfile content to Azure Cosmos Db from NiFi?
Thanks in advance
UPDATE(2019.05.26):
In the end I used Python script and called it from NiFi to post messages. I passed a message as a parameter. The reason I chose python is because it has some examples on official Microsoft site with all the required connection settings and libraries, so it was easy to connect to Cosmos.
I tried Mongo component, but couldn't connect to Azure (security config didn't work), didn't really go too far with it as Python script worked just fine.
Azure CosmosDB exposes MongoDB API so you can use the following MongoDB processors which are available in NiFi to read/query/write to & from Azure CosmosDB using Apache NiFi.
DeleteMongo
GetMongo
PutMongo
PutMongoRecord
RunMonogAggregation
Useful Links
https://learn.microsoft.com/en-us/azure/cosmos-db/mongodb-introduction
https://learn.microsoft.com/en-us/azure/cosmos-db/mongodb-feature-support
Valeria. According to the components list supported by Apache Nifi related to Azure, you could only get Azure Blob Storage, Queue Storage, Event Hub etc,not including Cosmos DB.
So,I suggest you using PutAzureBlobStorage to feed azure blob container with documents from NiFi flow files. Then please create a copy activity pipeline in Azure Data Factory to transfer data from Azure Blob Storage into Azure Cosmos DB.
Hello and I am new to developing with Qt for cross platform development. My current goal is to be able to cache data from a CosmosDB with my app that I've developed in Qt. I've had no problem setting up the CosmosDB, but I am confused at what the best way to communicate with the database is.
Do I need to create my own API to talk with the database? Are there libraries out there already do what I want?
You can use either standard HTTP requests, or use SQL API.
Azure Cosmos DB exposes resources through the REST APIs that can be called by HTTP/HTTPS requests. How can I develop apps with the SQL API
Azure Cosmos DB REST API reference
Also Azure Cosmos DB offers a query language as an interface to query JSON documents. The language supports a subset of ANSI SQL grammar and adds deep integration of JavaScript object, arrays, object construction, and function invocation. Microsoft shows exmaples here
I have a .NET web-app (4.5) the app is a cloud app in the sense that it is hosted remotely and has its own local SQL data source.
The web app needs to talk to client SQL servers (in this case lets say my local SQL database). At the core the web app will request data from the remote SQL source and save it to the local SQL source, or the other way around the web-app may notify the remote SQL source that it needs to update a particular record, etc.
What would be the best way to do this?
I was looking at just doing a windows service on the client SQL server, and the web app makes calls to the client SQL box web service telling it what it needs to update or send over?
I started looking into WCF would this be a good application of WCF?
You may want to consider implementing an OData service as a WCF Data Service hosted in IIS that exposes data queried from SQL Server.
The following MSDN article provides an overview of a SQL Server OData implementation:
http://msdn.microsoft.com/en-us/data/gg192995.aspx
The following CodePlex link seems to provide good and relevant information pertaining to your scenario:
http://msftdbprodsamples.codeplex.com/releases/view/101644
Regards,