Difference between revisions of "Certified HTTP"

From Second Life Wiki
Jump to navigation Jump to search
m (Reliable Message Delivery Over HTTP moved to Certified HTTP: I think this page's purpose has changed over time, it started out as a general discussion, and now it's a rudimentary spec for something specific.)
(Naming, naming, naming, naming! *does ballmer monkey dance*)
Line 1: Line 1:
AKA Certified HTTP.
= Goals =
= Goals =
The basic goal of certified http (colloquially known as chttp) is to have some kind of delivery and receipt guarantees for messages.
The basic goal of Certified HTTP (colloquially known as chttp) is to have some kind of delivery and receipt guarantees for messages.


From a standard http client perspective, if the client reads a whole response, then it knows for certain the server handled the request. However, for all other failure modes, the client can't be sure if the server did, or did-not perform the request function. On the server side, the server can never know if the client ever got the answer or not. For some operations, we need the ability for the client to perform a data-altering operation and be insistent that it occur. In particular, if it isn't certain that it happened, then it must be able to try again safely.
From a standard http client perspective, if the client reads a whole response, then it knows for certain the server handled the request. However, for all other failure modes, the client can't be sure if the server did, or did-not perform the request function. On the server side, the server can never know if the client ever got the answer or not. For some operations, we need the ability for the client to perform a data-altering operation and be insistent that it occur. In particular, if it isn't certain that it happened, then it must be able to try again safely.

Revision as of 16:34, 14 September 2007

Goals

The basic goal of Certified HTTP (colloquially known as chttp) is to have some kind of delivery and receipt guarantees for messages.

From a standard http client perspective, if the client reads a whole response, then it knows for certain the server handled the request. However, for all other failure modes, the client can't be sure if the server did, or did-not perform the request function. On the server side, the server can never know if the client ever got the answer or not. For some operations, we need the ability for the client to perform a data-altering operation and be insistent that it occur. In particular, if it isn't certain that it happened, then it must be able to try again safely.

The bigger picture goal is to make a simple way to conduct reliable delivery and receipt such that general adoption by the wider community is possible. This means that we have to respect HTTP where we can to take advantage of existing tools and methodology and to never contradict common conventions in a REST world.

competitive comparisons

I believe that all competitors to this fall into one of three categories, specialized message queues, reliable logic tunneled through HTTP, and delivery over HTTP with setup charges.

traditional message queue technology

This includes products like MQ, MSMQ, and ActiveMQ.

These technologies are useful, but provide a number of hurdles:

  • No common standards.
  • Integrates a new technology with unknown performance characteristics.
  • Requires significant operational overhead.

Because of this, we have opted to use HTTP as a foundation of the technology.

reliable application logic in body

This includes technologies like httpr and ws-reliable.

These tend to be thoroughly engineered protocol specifications which regrettably repeat the mistakes of the nearly defunct XMLRPC and the soon to join it SOAP -- namely, treating services as a function call. This is a reasonable approach, and is probably the most obvious to the engineers working on the problem. The most obvious path, which is followed in both of the examples, is to package a traditional message queue body into an HTTP body sent via POST. Treating web services as function calls severely limits the expressive nature of HTTP and should be avoided.

Consumers of web services prefer REST APIs rather than function calls over HTTP. I believe this is because REST is inherently more comprehensible to consumers since all of the data is data the consumer requested. In one telling data point, Amazon has both SOAP and REST interfaces to their web services, and 85% of their usage is of the REST interface[1]. I believe ceding power to the consumer is the only path to make wide adoption possible. In doing so, we must drop the entire concept of burying data the consumer wants inside application data.

reliable delivery after setup

There appear to be a few proposals of this nature around the web and a shining example can be found in httplr.

These are http reliability mechanisms which require a round trip to begin the reliable transfer and then follow through with some permutation of acknowledging the transfer. This setup can be cheap since the setup does not have the same reliability constraints as long as all clients correctly handle 404. Also, some of this overhead can be optimized away by batching the setup requests.

However, a protocol which requires setup for all messaging will always be more expensive than other options.

Requirements

We have the concept of a "long" time (LT). The purpose of this time is to ensure that any clients will stop trying a particular request long before the server discards state associated with the request. This is not a very theoretically "pure" concept, since it relies on all parties conforming to certain (reasonable) assumptions, but we believe that we get significant benefits (lack of explicit negotiation), at little risk of actually entering an incorrect state. chttp is expected to generate a small amount of bookkeeping data for each message, and there needs to be a way to expire this data in a way that doesn't involve additional negotiation between servers and clients. The server should keep data for LT, and clients should not retry messages that are older than LT/2 or similar, so that there is at least LT/2 for a Reliable Host to recover from errors, and LT/2 for clock skew. LT therefore should be orders of magnitude longer than the longest downtime we expect to see in the system and the largest clock skew we expect to see. Off the cuff, 30 days seems like a reasonable value for LT.

  • chttp is based on http, and can use any facility provided by http 1.1 where not otherwise contradicted.
    • this includes the use of https, pipelining, chunked encoding, proxies, redirects, caches, and headers such as Accept, Accept-Encoding, Expect, and Content-Type.
    • any normal http verb appropriate to context should be accepted, eg POST, PUT, GET, DELETE
    • unless otherwise specified, the http feature set in use is orthogonal and effectively transparent to chttp
  • in any complete chttp exchange the client and server can eventually agree on success or failure of delivery, though it is more important that no ill effects arise when they disagree
  • any message will be effectively received once and only once or not at all
  • the content of the http body must be opaque to chttp
  • the URI of the original request must be opaque to chttp
  • chttp enabled clients and servers can integrate with unreliable tools
    • the chttp server can differentiate reliable requests and respond without reliability guarantees (i.e. act as a normal http server)
    • chttp clients can differentiate reliable responses and handle unreliable servers (i.e. act as a normal http client)
  • the client will persist the local time of sending
  • if there is one the client must either have the persisted outgoing body or the exact same body can be regenerated on the fly
  • the server will persist the local time of message receipt
  • the server must persist the response body or have a mechanism to idempotently generate the same response to the same request
  • all urls with a chttp server behind them are effectively idempotent for all uniquely identified messages
    • the server will keep receipt of messages for a minimum of LT days
    • the client can retry steadily over a period of LT/2 days
    • the client and server are assumed to almost always be running
    • over that window of opportunity, the same message will always get the same response
  • all persisted data on a single host is ACID

requirements on top of http

  • if the body of the request is non-zero length, the client MUST include a Content-Length header, unless prohibited by section 4.4 of RFC 2616
    • the server will look for \r\n\r\n and content length body to consider the request complete
    • an incomplete request will result in 4XX status code
  • if the body of the response is non-zero length, the server must include a Content-Length header
    • the client will look for \r\n\r\n and content length body to consider the response complete
    • the client will retry on an incomplete response
  • Messages can not be terminated by closing the connection -- only positive expressions of message termination (such as Content-Length) can guarantee that a full message is received.

assumptions

  • The client and server will not have any significant time discontinuity, i.e., the clock difference between them should be less than LT/100.
  • The client and server will not have clock drift more than LT/100 per day.
  • Both parties exchanging messages are Reliable Hosts (defined below).
  • Generating a globally unique message ID is inexpensive.
  • It is possible to store a large amount of 'small' data (such as UUIDs, urls, and date/timestamps) for LT. "Large" in this case means "as many items as the throughput of a host implies that you could create during LT". In other words, storing this small data will never be a resource problem.
  • Any transaction or data handled by this system will become useless well before LT has elapsed.
Reliable Host
A reliable host has the following properties:
  • The host has a durable data store of finite size, which can store data in such a way that it is guaranteed to be recoverable even in the face of a certain number of hardware failures.
  • The host will not be down forever. Either a clone will be brought up on different hardware, or the machine itself will reappear within a day or so.
  • The host can perform "Wikipedia logo"ACID operations on the data it contains.

Implementation

Requiring an opaque body and URI pretty much requires either negotiating an URL beforehand or adding new headers. Since the former was discarded earlier (reliable delivery after setup), we will focus on adding request and response headers.

This is a suggested implementation:

  1. chttp enabled servers will behave idempotently with a unique message id on the request
  2. the server can request acknowledgment from the client if the request consumes server resources

Request

A client will start a reliable message with a GET, POST, or PUT to an url on a reliable http server. The url may be known in advance or returned as part of an earlier application protocol exchange. The message itself will be uniquely identified by a new header for the message id.

X-Message-ID

A combination of client host identifier with a uuid or sequence number which specifies the reliability client ID for the request. When a server detects the X-Message-ID, it can guarantee reliable message delivery.

Since the message id must be unique, implementations should generate an id of the form $random_uuid@$client_host where the uuid is from a cryptographically secure generator and client_host is the name of the host.

Sending a message with the same message id but a a different body has an undefined result.

Response

Generally, when the client gets the 2XX back from the server, the message has been delivered. If the server has a response body, then the client will need to acknowledge receipt of the entire body. If the entire body is not read, the client can safely resubmit the exact same request. The server will include a message url in the headers if the client specified a message id on request and the body requires persistence by the server.

X-Message-URL

This will be a unique url on the server for the client message id which the client will DELETE upon receipt of the entire response to the request. This will only exist if there is a non-null response body or the server is consuming significant resources to maintain that url. When the client performs the delete, the server can flush all persisted resources other than the fact that the message was sent at some point.

If the client performs a GET to the message url prior to DELETE, the server must return the same body as the original response, including the X-Message-URL.

Workflow

The workflow is a specification/pseudocode for what actions both ends of chttp communication need to take to fulfill the requirements.

Sending a Message

The sending-side API looks very much like a normal HTTP method call:

 response = certified_http.put(url, body)

What happens under the covers is:

  1. Generate a globally unique message ID for the message
  2. Store the outgoing request (headers and all), including the message id, in a durable store "outbox", and waits for a response.
  3. A potentially asynchronous process performs the following steps:
    1. Retrieves the request from the outbox.
    2. Performs the HTTP request specified by the outbox request and waits for a response.
    3. If a response is not forthcoming, for whatever reason, the process retries after a certain period.
    4. If the server sends an error code that indicates that the reliable message will never complete (e.g. 501), or a long timeout expires indicating that an absurd amount of time has elapsed, the method throws an exception.
  4. Opens a transaction on the durable store
  5. Stores the incoming response in a durable inbox.
  6. 'Tombstones' the message in the outbox, which essentially marks the message as having been received, so that if the application resumes again, it doesn't resend.
  7. Closes the transaction on the durable store
  8. If the response contains a header indicating a confirmation url on the recipient, performs an HTTP DELETE on the resource to ack the incoming message.

There are no explicit semantics for the response body, like HTTP itself. The content will vary depending on the application.

Receiving a Message

The receiver sets up a node in the url hierarchy, just like a regular http node. When an incoming request comes in, the receiver:

  1. Stores the incoming request in a durable store "inbox", if it doesn't already contain a message with the same ID.
  2. A potentially asynchronous process performs the following steps:
    1. Looks for responses in the outbox matching the incoming message id, and if it finds one, sends it as the response without invoking anything else.
    2. Opens a transaction on the database, locking the inbox request
    3. Calls the handler method on the receiving node:
      1. def handle_put(body, txn):
        return "My Response"
      2. The handler method can use the open transaction to perform actions in the database that are atomic with the receipt of the message. Any non-idempotent operation must be done atomically in this way.
    4. Stores the return value of the handle method as an outgoing response in the outbox, without closing the transaction.
    5. Removes the incoming request from the inbox
    6. Closes the transaction
  3. Discovers a new item in the outbox, responds to the incoming http request with the response from the outbox, including a url that, if DELETEd, will remove the item from the outbox.

Performing a Job/RPC

There is no special logic for performing a job (which is another way of saying "an RPC that may take a while"). If a long-running job is to be performed, the Receiver simply delays its response until the job is complete. This may result in the Sender timing out and retrying, but that's OK because the Receiver will simply respond to whichever retry happens to occur at or after the time the response is put in the outbox.