background preloader

Hypertext Transfer Protocol

Hypertext Transfer Protocol
The Hypertext Transfer Protocol (HTTP) is an application protocol for distributed, collaborative, hypermedia information systems.[1] HTTP is the foundation of data communication for the World Wide Web. The standards development of HTTP was coordinated by the Internet Engineering Task Force (IETF) and the World Wide Web Consortium (W3C), culminating in the publication of a series of Requests for Comments (RFCs), most notably RFC 2616 (June 1999), which defined HTTP/1.1, the version of HTTP most commonly used today. In June 2014, RFC 2616 was retired and HTTP/1.1 was redefined by RFCs 7230, 7231, 7232, 7233, 7234, and 7235.[2] HTTP/2 is currently in draft form. Technical overview[edit] URL beginning with the HTTP scheme and the WWW domain name label. A web browser is an example of a user agent (UA). HTTP is designed to permit intermediate network elements to improve or enable communications between clients and servers. History[edit] The first documented version of HTTP was HTTP V0.9 (1991).

Internet Protocol This article is about the IP network protocol only. For Internet architecture or other protocols, see Internet protocol suite. The Internet Protocol (IP) is the principal communications protocol in the Internet protocol suite for relaying datagrams across network boundaries. Its routing function enables internetworking, and essentially establishes the Internet. Historically, IP was the connectionless datagram service in the original Transmission Control Program introduced by Vint Cerf and Bob Kahn in 1974; the other being the connection-oriented Transmission Control Protocol (TCP). The first major version of IP, Internet Protocol Version 4 (IPv4), is the dominant protocol of the Internet. Function[edit] The Internet Protocol is responsible for addressing hosts and for routing datagrams (packets) from a source host to a destination host across one or more IP networks. Datagram construction[edit] Sample encapsulation of application data from UDP to a Link protocol frame Reliability[edit]

HTTP Secure Hypertext Transfer Protocol Secure (HTTPS) is a communications protocol for secure communication over a computer network, with especially wide deployment on the Internet. Technically, it is not a protocol in and of itself; rather, it is the result of simply layering the Hypertext Transfer Protocol (HTTP) on top of the SSL/TLS protocol, thus adding the security capabilities of SSL/TLS to standard HTTP communications. The security of HTTPS is therefore that of the underlying TLS, which uses long term public and secret keys to exchange a short term session key to encrypt the data flow between client and server. To guarantee one is talking to the partner one wants to talk to, X.509 certificates are used. In its popular deployment on the internet, HTTPS provides authentication of the web site and associated web server that one is communicating with, which protects against man-in-the-middle attacks. Overview[edit] HTTPS creates a secure channel over an insecure network. Usage in websites[edit]

S-T: Http Request/Response Anatomy Introduction to Web Web consist of billion of clients and server connected through wires and wireless networks. The web clients makes request to web server. The web server receives the request, finds the resources and return response to the client. Web Application A web site is a collection of static files such as HTML pages, images, graphics etc. HTTP is a protocol that clients and server uses on the web to communicate.It is similar to other internet protocol such as SMTP(Simple Mail Transfer Protocol) and FTP(File Transfer Protocol) but there is one fundamental difference.HTTP is a stateless protocol i.e HTTP supports only one request per connection.This means that with HTTP the clients connects to the server to send one request and then disconnects. HTTP method HTTP request can be made using a variety of methods, but the ones you will use most often are Get and Post. HTTP Method and Descriptions Difference between GET and POST request Anatomy of an HTTP GET request

PHP Scripting language created in 1994 PHP is a general-purpose scripting language geared towards web development.[8] It was originally created by Danish-Canadian programmer Rasmus Lerdorf in 1993 and released in 1995.[9][10] The PHP reference implementation is now produced by the PHP Group.[11] PHP was originally an abbreviation of Personal Home Page,[12][13] but it now stands for the recursive initialism PHP: Hypertext Preprocessor.[14] The standard PHP interpreter, powered by the Zend Engine, is free software released under the PHP License. The PHP language has evolved without a written formal specification or standard, with the original implementation acting as the de facto standard that other implementations aimed to follow. History[edit] Early history[edit] PHP development began in 1993[9] when Rasmus Lerdorf wrote several Common Gateway Interface (CGI) programs in C,[20][21] which he used to maintain his personal homepage. <! PHP 3 and 4[edit] PHP 5[edit] PHP 6 and Unicode[edit] PHP 7[edit]

Magnet URI scheme Scheme that defines the format of magnet links Magnet is a URI scheme that defines the format of magnet links, a de facto standard for identifying files (URN) by their content, via cryptographic hash value rather than by their location. Although magnet links can be used in a number of contexts, they are particularly useful in peer-to-peer file sharing networks because they allow resources to be referred to without the need for a continuously available host, and can be generated by anyone who already has the file, without the need for a central authority to issue them. This makes them popular for use as "guaranteed" search terms within the file sharing community where anyone can distribute a magnet link to ensure that the resource retrieved by that link is the one intended, regardless of how it is retrieved. History[edit] Format [edit] The following parameters are supported:[2][3] The standard also allows for application-specific experimental parameters, which must begin with "x". magnet:? x.

HTTP: The Protocol Every Web Developer Must Know - Part 2 In my previous article, we covered some of HTTP's basics, such as the URL scheme, status codes and request/response headers. With that as our foundation, we will look at the finer aspects of HTTP, like connection handling, authentication and HTTP caching. These topics are fairly extensive, but we'll cover the most important bits. A connection must be established between the client and server before they can communicate with each other, and HTTP uses the reliable TCP transport protocol to make this connection. HTTPS is a secure version of HTTP, inserting an additional layer between HTTP and TCP called TLS or SSL (Transport Layer Security or Secure Sockets Layer, respectively). An HTTP connection is identified by <source-IP, source-port> and <destination-IP, destination-port>. resolve IP address from host name via DNSestablish a connection with the serversend a requestwait for a responseclose connection The server is responsible for always responding with the correct headers and responses.

File Transfer Protocol FTP is built on a client-server architecture and uses separate control and data connections between the client and the server.[1] FTP users may authenticate themselves using a clear-text sign-in protocol, normally in the form of a username and password, but can connect anonymously if the server is configured to allow it. For secure transmission that hides (encrypts) the username and password, and encrypts the content, FTP is often secured with SSL/TLS ("FTPS"). SSH File Transfer Protocol ("SFTP") is sometimes also used instead, but is technologically different. History[edit] The original specification for the File Transfer Protocol was written by Abhay Bhushan and published as RFC 114 on 16 April 1971. Until 1980, FTP ran on NCP, the predecessor of TCP/IP.[2] The protocol was later replaced by a TCP/IP version, RFC 765 (June 1980) and RFC 959 (October 1985), the current specification. Protocol overview[edit] Communication and data transfer[edit] ASCII mode: used for text. Login[edit] or:

BitTorrent Programmer Bram Cohen, a former University at Buffalo graduate student in Computer Science,[4] designed the protocol in April 2001 and released the first available version on July 2, 2001,[5] and the final version in 2008.[6] BitTorrent clients are available for a variety of computing platforms and operating systems including an official client released by Bittorrent, Inc. As of 2009, BitTorrent reportedly had about the same number of active users online as viewers of YouTube and Facebook combined.[7][8] As of January 2012[update], BitTorrent is utilized by 150 million active users (according to BitTorrent, Inc.). Based on this figure, the total number of monthly BitTorrent users can be estimated at more than a quarter of a billion.[9] Description[edit] The middle computer is acting as a seed to provide a file to the other computers which act as peers. The file being distributed is divided into segments called pieces. When a peer completely downloads a file, it becomes an additional seed.

Cascading Style Sheets CSS is designed primarily to enable the separation of document content from document presentation, including elements such as the layout, colors, and fonts.[1] This separation can improve content accessibility, provide more flexibility and control in the specification of presentation characteristics, enable multiple pages to share formatting, and reduce complexity and repetition in the structural content (such as by allowing for tableless web design). CSS can also allow the same markup page to be presented in different styles for different rendering methods, such as on-screen, in print, by voice (when read out by a speech-based browser or screen reader) and on Braille-based, tactile devices. It can also be used to allow the web page to display differently depending on the screen size or device on which it is being viewed. CSS specifies a priority scheme to determine which style rules apply if more than one rule matches against a particular element. Syntax[edit] Selector[edit] Use[edit]

eDonkey network The eDonkey Network (also known as the eDonkey2000 network or eD2k) is a decentralized, mostly server-based, peer-to-peer file sharing network best suited to share big files among users, and to provide long term availability of files. Like most sharing networks, it is decentralized, as there is not any central hub for the network; also, files are not stored on a central server but are exchanged directly between users based on the peer-to-peer principle. Currently, the eD2k network is not supported by any organization (in the past it was supported by the MetaMachine Corporation, its creator, which now is out of business) and development and maintenance is being fully provided by its community and client developers. There are many programs that act as the client part of the network. The original eD2k protocol has been extended by subsequent releases of both eserver and eMule programs, generally working together to decide what new features the eD2k protocol should support. Features[edit]

Web Services Description Language The Web Services Description Language (WSDL) is an XML-based interface description language that is used for describing the functionality offered by a web service. The acronym is also used for any specific WSDL description of a web service (also referred to as a WSDL file), which provides a machine-readable description of how the service can be called, what parameters it expects, and what data structures it returns. It thus serves a purpose that corresponds roughly to that of a method signature in a programming language. The current version of WSDL is WSDL 2.0. The meaning of the acronym has changed from version 1.1 where the D stood for Definition. Description[edit] Representation of concepts defined by WSDL 1.1 and WSDL 2.0 documents. The WSDL describes services as collections of network endpoints, or ports. Example WSDL file[edit] <? History[edit] WSDL 1.0 (Sept. 2000) was developed by IBM, Microsoft, and Ariba to describe Web Services for their SOAP toolkit. See also[edit] References[edit]

ed2k URI scheme In computing, eD2k links ( are hyperlinks used to denote files stored on computers connected to the eDonkey filesharing P2P network. General[edit] Many programs, such as eMule, MLDonkey and the original eDonkey2000 client by MetaMachine, which introduced the link type, as well as others using the eDonkey file sharing protocol, can be used to manage files stored in the filesharing network. eD2k links allow a file to be identified from a link in a web browser and to be downloaded thereafter by a client like eMule, Shareaza or any other compatible software. This linking feature was one of the first URIs to be introduced in peer-to-peer file sharing, and had a vast effect on the development of the eDonkey network, as it allowed external link sites to provide verified content within the network. Like other URI protocols, web browsers can be configured to automatically handle ed2k URIs. File link format[edit] eD2k hash algorithm[edit] Server links[edit] Example:

Related: