Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 7 years ago.
Improve this question
First of all, this is obviously for learning porpoises only. Don't be afraid to answer.
So anyway, is there a way for a man to hack a computer logging in, for example, to a garry's mod server hosted on my PC?
If there is, I am very interested on how this is working. Explanations will be welcome.
In general terms, yes, it's possible. Game clients receive data from their servers, which they expect to be in a particular format. If the server is modified to send mis-formatted data, the result could easily be to trigger a buffer overflow or other exploitable bug in the client.
See for example http://threatpost.com/researchers-discover-dozens-of-gaming-client-and-server-vulnerabilities/100744
Not sure about your locale, but most countries have a similar law to the UK's Computer Misuse Act. Which pretty much means 'hacking back' is illegal.
If you want to learn about exploits, and how to use them ethically - www.google.com is the place to start, try looking for 'ethical hacking course'
Related
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 4 years ago.
Improve this question
I'm building a program that I want to be able to exchange information with other programs running in another computer. I started using C# and a library called SimpleTCP. The main issue is that is too simple and only send and receive messages.
I'm looking for something that I can predetermine functions that one or the other can call from each other.
I looked on google and stack overflow but I was unable to find an appropriated subject to study, what should I be looking for to learn this? Thank you
The most complete protocol for what you want is gRPC. There is a learning curve but worth it in my opinion. https://github.com/grpc/grpc
There is a way but it's little bit different
Such programs like this written in tow different languages
You can make a center database between the both programs
In this situation it's very easy to communicate and receive ,send data
You can mysql ,oracl, mariadb or any Database you prefer
Closed. This question is opinion-based. It is not currently accepting answers.
Want to improve this question? Update the question so it can be answered with facts and citations by editing this post.
Closed 7 years ago.
Improve this question
I’m considering buying HTTP: The Definitive Guide with the goal of properly understanding HTTP and related technologies. I’m wondering if—13 years after its publication—the book is still worth buying.
To help me make a decision, I’m looking for input on two questions:
What important changes have happened in the world of HTTP that the book doesn’t cover? (For bonus points: what resources are recommended reading to catch up?)
Are any of the sections not relevant anymore at all?
Check the book’s table of contents to get an idea of what’s covered.
Edit: This book isn’t about web fronted technologies. Obviously, there have been dramatic changes in that area since 2002. Please answer with respect to HTTP. Thanks.
HTTP 1.1 has been around since the late 90s. The information should still be very relevant.
http://www.w3.org/Protocols/History.html
HTTP 2.0 is starting to gain momentum. It has been majorly revised and is a binary protocol now. But there are probably not many books written about it at this time.
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I have been asked to design a website for a client as a "side job". I am trying to write up a statement of work for the project. In the past, I have done similar work, and often run into a situation where I believe the work is "done", but the client wants endless tweaks and changes. (As you know, websites are perpetually "under construction").
When you have requirements such as "Design a Home page, design a Contact Us page" how you define a page as "done"?
Don't put anything live, until they accept your work is complete. This should be enough of an incentive for them not to string you along, and allows them to have the quality website they require.
Ask the client to set up a requirements specification for version 1. When you met the requirements contained in this document is your job completed. Everything else belongs to the next version.
In the same situations, I tell my client "you want A, B, C and D. OK, sign here, and we are agreed that the end of application is A-D. Now if you wanted something more in future, it is not a part of our contract, so we'll deal with that in future and of course it has it's own price." This way you make them think before signing and lot's of things become clearer, and lots of needs show up suddenly, but in future they'll either pay more for more needs or won't talk any more :)
Closed. This question is off-topic. It is not currently accepting answers.
Want to improve this question? Update the question so it's on-topic for Stack Overflow.
Closed 10 years ago.
Improve this question
I have a project idea for which I want to mine publicly available data on another website that it received by crowd-sourcing. This is so I have initial data for my own project. To reiterate, I want to write a robot to grab data that is displayed on another website and use it for my own website. Does anyone know the legality of this sort of thing? Does the original website own the data that was given to it by a crowd? Even if so, can I use it?
Web scraping is a legally complicated issue.
The hassles of legal action and enforceability often keep scrapers from getting in trouble.
Outright duplication is considered actionable, although courts have ruled that "duplication of facts" is permitted (US).
I advise you read up here: http://en.wikipedia.org/wiki/Web_scraping#Legal_issues
Best,
legally, you should be fine. as long as the data is made available and the people have consented; you aren't hacking and the other site has permission to share. check for a license on the other site, if there isn't one inquire or be prepared for access to be denied at some point. and even though it is publicly available doesn't mean the other site wants it to be.
also, double check and make sure that you don't inadvertently publish private data as well.
Closed. This question needs to be more focused. It is not currently accepting answers.
Want to improve this question? Update the question so it focuses on one problem only by editing this post.
Closed 6 years ago.
Improve this question
My software is a video-audio converter and video cutter. I have used Qt(compiled from source) and ffmpeg (compiled from source). I have to prepare System Usage Specification outline and Specify Usage patterns of the system and indicate it using Run charts / Histograms. I am told to use Winrunner for this purpose. I don't know exactly what to do. Please help.
I never heard about 'System Usage Specification', this must be a terminology specific to your company.
A wild guess would be that it's something close to the Use Case diagram of UML, to define what the users can do and which action they have to perform to lead them to the expected result.
Sounds like a uggly word for "handbook" or "usage guide" from the pov of a end-user (though I never heared of that specific term)