Closed. This question needs details or clarity. It is not currently accepting answers.
Want to improve this question? Add details and clarify the problem by editing this post.
Closed 4 years ago.
Improve this question
I am making a small project where there is a qt client application and basic server application written in c++. I am able to connect between two different computers through my application (over SSL, Socket). Now I want to connect my client application to access a sqlite database which is on server.
For example. lets say my client application is a simple login form. When user press Login button then I want my application to access the sqlite database on server, and verify whether the details are valid.
Above a just a example fr explanation. I know I can do the above thing in various other ways but I want the above functionality for something different.I searched a lot over internet but unable to find a good example or explanation. Please through some insight.
SQLite data is stored in a single file, and there's no database server running on top of it. So it does not have any network connection capability. A very simple solution is to make the SQLite database file available on a network in some way like sharing it's directory.
There are other options which use third party libraries :
netSQLite : a client/server solution running over TCP/IP with SQLite3 at its core.
Navicat : enables remote SQLite connection although it is not free.
SQLiteServer: a multithreaded TCP/IP server for SQLite
SQLiteDBMS: a database management server for SQLite. It allows an sqlite3 process to be accessed via a TCP/IP network.
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 3 years ago.
Improve this question
SGX remote attestation procedure (variations: with and without the participation of the remote Intel service, the explanation / logic of the work is step by step.
Sadly, remote attestation is not possible without the Intel Attestation Service, which I'll call IAS from now on. Also, the party that's want to verify a quote needs to be registered as service provider (short SP) to the IAS. I'll now start with a high level overview of the typical of the remote attestation procedure:
The enclave connects to the SP and sends it's attestation group id.
The SP fetches a list with revoked clients, also known as SigRL, from the IAS and sends it to the client.
The client creates a quote using the SigRL.
The SP forwards the quote to the IAS which responds whether the quote is valid
On a lower level, the client will have to establish communication between the quoting enclave (short QE) which is provided by Intel and the enclave that want to remotely attest something, which usually is done using the AESM-service (which is part of the platform software).
Also, the quoting mechanism allows you to remotely attest some data which you should use since it makes the remote attestation proof that this particular data was generated in an enclave which enables trusted protocols. There is an extensive examle which does a key-exchange on Intels Devzone.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
Hi guys I'm creating a device that can be controlled by google assistant;
currently I created an arduino that interfaces with an esp8266 and through the service "IFTTT" I can communicate without problems. Now I would like to extend the project so as to make it public so as to find it in the "Home control" devices of google.
Is it possible to use "Arduino / esp8266"? How?
If it is not possible in this way what kind of board do you recommend?
In order for your device to work with the Google Assistant, you should consult the smart home documentation. If your device can connect to the Internet, you're good. The ESP8266, since it already connects through IFTTT, will work for a more public project as well.
You'll need to create some sort of server backend which will act as the source of truth, responding to requests from the Google Assistant and dispatching commands to devices. If you can use the ESP8266 as a web server with minimal (< 5s) latency, you'll be set. Alternatively, you can use a simple service like Firebase.
Your web server will need to respond to three primary types of intents: SYNC, QUERY, and EXECUTE. The documentation walks through all of them, but they give you the ability to provide devices for each user, return the state of these devices, and send commands to do actions.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
We are planning to develop an ERP-project using ASP.NET and SQL. Client asked one feature: to be able to access and use ERP even if connection to Internet fails (using without Internet). Is it possible?
Of course you can use that without internet. You may host your web-site on some server in that company and all employers of that company may use this ERP. But nobody outside company will not be able to connect to that. Government committees usually use this way. If they scare about privacy and security, they may just take VPN access, it is secure. Another way (the worst, as I think): users from outside connect to computers inside company, and work over that computers.
Second solution: Web-based asynchronous application. You will create something like framework, you will install that to all users' computer. What is the principle of that's work? All user's will have their own database file, and they will work with their database. You just need to make a synchronization service between databases. For example, every time when user have internet, your service will synchronize data between user's database and server's database.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
The community reviewed whether to reopen this question 10 months ago and left it closed:
Original close reason(s) were not resolved
Improve this question
I'm trying to get a deeper understanding of how IIS works.
http.sys i understand is one its major components. However, i have been having trouble finding easily digestible information about it. I couldn't get a good mental model going until i heard about the WSK, then i think it all fell into place.
From a lot of random googling a little experimentation this is my current high level understanding of why it exists and how it does it's stuff.
Why:
Port sharing, and higher performance caching.
How:
User mode processes use the WinSock api to open a socket listening on a port to gain access to the networking subsystem, e.g. tcp/ip. Kernal mode software like the http.sys driver uses Winsock Kernal Sockets (WSK) api to achieve the same end using the same pool of TCP port numbers as the WinSock api.
IIS, a web service or anything that wants to use http registers itself with http.sys using a unique url/port combination. http.sys opens up a socket on this port using WSK (if it hasn't already for another url/port combination with the same port) and listens.
When the transport layer (tcpip.sys) has reconstructed a load of ip packets back into an http request that a client sent it gives it to http.sys via the port in the request. Http.sys uses the url/port number to send it the the appropriate process which parses it however it pleases.
I know it seems like I'm answering my own question but I'm really not that sure of myself on this and would like some closure so i can get on with more interesting things.
Am i close?
Closed. This question does not meet Stack Overflow guidelines. It is not currently accepting answers.
We don’t allow questions seeking recommendations for books, tools, software libraries, and more. You can edit the question so it can be answered with facts and citations.
Closed 6 years ago.
Improve this question
I was wondering where I could learn more about decentralized sharing and P2P networks. Ideally, I'd like to create something to help students share files with one another over their universities network, so they could share without fear of outside entities.
I'm not trying to build the next Napster here, just wondering if this idea is feasible. Are there any open source P2P networks out there that could be tweaked to do what I want?
Basically you need a server (well, you don't NEED a server, but it would make it much simplier) that would store user IPs between other things like file hash lists, etc.
That server can be in any enviroinment you want (which is very comfortable).
Then, each client connects to the server (it should have a dns, it can be a free one, I've used no-ip.com once) and sends basic information first (such as its IP, and a file hash list), then sends something every now and then (say each 5 minutes or less) to report that it's still reachable.
When a client searchs files/users, it just asks the server.
This is a centralized network, but the file sharing would be done in p2p client-to-client connections.
The reason to do it like this is that you can't know an IP to connect to without some reference.
Just to clear this server thing up:
- Torrents use trackers.
- eMule's ED2K uses lugdunum servers.
- eMule's "true p2p" Kademlia uses known nodes (clients) (most of the time taken from servers like this).
Tribler is what you are looking for!
It's a fully decentralized BitTorrent Client from the Delft University of Technology. It's Open Source and written in Python, so also a great starting point to learn.
Use DC++
What is wrong with Bit-Torrent?
Edit: There is also a pre-built P2P network on Microsoft operating systems that is pretty cool as the basis to build something. http://technet.microsoft.com/en-us/network/bb545868.aspx