As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
Does anybody know of an implementation of TCP on a FPGA WITHOUT using any sort of microblaze? Preferably open source, because it is for an university/research project.
Depending on what you want you maybe can get away with a relative small own implentation (e.g. for packet inspection). The statefulness of TCP makes an full hardware implementation vary big and cumbersome. If possible I would recommend to switch to UDP, that makes it much easier.
As project dealing with all the IP stuff I know NetFPGA, but I never checked their design, so it could be, that they utilize internal a microblaze for some stuff, but my guess would be not.
EDIT: I also remember, that I met one someone from the University of Copenhagen (not sure about this point) at a conference, who also implemented TCP stack on Xilinx FPGAs.
I know Easics has a TCP core. You can find a presentation on it here
As far as I know both Intelop and Velocytech have commercial TCP/IP cores available
A full and low latency TCP/IP Hardware Stack is also available at PLDA
Related
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
As written in the heading, my question is, why does TCP/IP use big endian encoding when transmitting data and not the alternative little-endian scheme?
RFC1700 stated it must be so. (and defined network byte order as big-endian).
The convention in the documentation of Internet Protocols is to
express numbers in decimal and to picture data in "big-endian" order
[COHEN]. That is, fields are described left to right, with the most
significant octet on the left and the least significant octet on the
right.
The reference they make is to
On Holy Wars and a Plea for Peace
Cohen, D.
Computer
The abstract can be found at IEN-137 or on this IEEE page.
Summary:
Which way is chosen does not make too much
difference. It is more important to agree upon an order than which
order is agreed upon.
It concludes that both big-endian and little-endian schemes could've been possible. There is no better/worse scheme, and either can be used in place of the other as long as it is consistent all across the system/protocol.
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
We are going to develope one web application using Asp.Net which can have millions of data to handle
so i am confuse between database selection
which should i prefer sql server or oracle with respect to performance and all criteria
please guide me on this
thanks
Your question is looks subjective, how ever I like to answer and say that:
If some one gives you to drive a formula one, in how many seconds you gong to crash it? Probably you do not even manage to start it running.
The same think is on programming. Both programs are like formula one, maybe one have some feature and the other have some other, but they can run so fast if "you can drive them" like that.
Now it's up to you to make a good design to the database and make it real fast, or very slow and huge. It's not the machine, it you that you can make it run fast. It's not the formula one on the races, it’s the pilot (and the rest team) that they drive them so fast.
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 11 years ago.
There are a lot of opensources build systems, and most of them are multiplatform, multilanguage, cross-everything.
I'm not interested on which one is the best (which would be offtopic by the way), but I'd like to know which ones are used most.
So, do we have any kind of statistic about their actual usage? Could you provide any link?
As there is no tracking process, I doubt that this kind of information is available. You could check the number of artifacts in maven central for example, but it won't tell that much about actual usage and may be viewed rather as a trend. Some build tool developers list some of the most reknowned users on their sites, most don't
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 11 years ago.
More of a side thing I want to learn since they never really went much over it in college, and yes im sure this has been asked plenty of times but Books/Algorithms get updated. So I wanted to most updated opinions/reviews of it.
I am less concerned about the history of it, but more concerned about actual implementation. And maybe by the end of the book implementing my own pseudo encryption algorithm.
I recall hearing something about Crpytool being a good learning program. I don't know if that is still true or not.
I am guessing typical encryption algorithms can be pretty much implemented in most languages right? Like MD5 in php and c?
http://www.schneier.com/book-practical.html
And please don't use your pseudo-encryption algorithm for anything more important than your own love letters. It's probably best not to use your own implementations of standard algorithms either.
Here is Cryptography for Developers and Cryptography in C and C++ books which I advise you much
The handbook of applied cryptography is worth to read, especially chapter 14.
http://www.cacr.math.uwaterloo.ca/hac/
Wikipedia has a good bibliography page on the subject which history page shows frequent updates.
As it currently stands, this question is not a good fit for our Q&A format. We expect answers to be supported by facts, references, or expertise, but this question will likely solicit debate, arguments, polling, or extended discussion. If you feel that this question can be improved and possibly reopened, visit the help center for guidance.
Closed 10 years ago.
Every day I write Web applications, and I have a good understanding of HTTP. However I want to close the gaps in my knowlege of network architecture. I'm not a sysadmin, so a hard-core sysadmin reference book would probably be a bit much for me, but I'm also not looking for a book on how to write code in any way -- I'm interested in the mechanisms underneath all that fun Web code I write.
Any recommendations?
TCP/IP Illustrated, Volume 1: The Protocols by W. Richard Stevens is the book you want to read.
http://www.amazon.com/TCP-Illustrated-Vol-Addison-Wesley-Professional/dp/0201633469
Also by the great W. Richard Stevens is Unix Network Programming (http://www.amazon.com/UNIX-Network-Programming-Networking-Sockets/dp/013490012X).