W3C

Web Architecture

Web Architecture focuses on the foundation technologies and principles which sustain the Web, including URIs and HTTP.

Architecture Principles Header link

Web Architecture principles help to design technologies by providing guidance and articulating the issues around some specific choices.

Identifiers Header link

We share things by their names. URL, URI, IRI is the way to name things on the Web and manipulate them. Some additional addressing needs in the Web Services stack motivated some additional layers.

Protocols Header link

Protocols are the vehicle for exchanging our ideas. HTTP is the core protocol of the Web. W3C is also working on XML Protocols and SOAP in relation to Web Services.

Meta Formats Header link

XML, the Extensible Markup Language, is used to build new formats at low cost (due to widely available tools to manipulate content in those new formats). RDF and OWL allow people to define vocabularies (“ontologies”) of terms as part of the Semantic Web.

Protocol and Meta Format Considerations Header link

Documents on the Web are loosely joined pieces by identifiers. It creates a maze of rich interactions between protocols and formats.

Internationalization Header link

W3C has worked with the community on the internationalization of identifiers (IRIs) and a general character model for the Web.

News Atom

The Encoding specificationhas been published as a Candidate Recommendation. This is a snapshot of the WHATWG document, as of 4 September 2014, published after discussion with the WHATWG editors. No changes have been made in the body of this document other than to align with W3C house styles. The primary reason that W3C is publishing this document is so that HTML5 and other specifications may normatively refer to a stable W3C Recommendation.

Going forward, the Internationalization Working Group expects to receive more comments in the form of implementation feedback and test cases. The Working Group
believes it will have satisfied its implementation criteria no earlier than 16 March 2015. If you would like to contribute test cases or information about implementations, please send mail to www-international@w3.org.

The utf-8 encoding is the most appropriate encoding for interchange of Unicode, the universal coded character set. Therefore for new protocols and formats, as well as existing formats deployed in new contexts, this specification requires (and defines) the utf-8 encoding.

The other (legacy) encodings have been defined to some extent in the past. However, user agents have not always implemented them in the same way, have not always used the same labels, and often differ in dealing with undefined and former proprietary areas of encodings. This specification addresses those gaps so that new user agents do not have to reverse engineer encoding implementations and existing user agents can converge.

The XML Localization Interchange File Format (XLIFF) version 2.0 has been approved as an OASIS Standard.

XLIFF is the open standard bi-text format: Bi-text keeps source language and target language data in sync during localization.

The publication of XLIFF 2.0 is of high importance for W3C since several of the main ITS 2.0data categories can be used within XLIFF 2.0 to provide content related information during the localization process. Full ITS 2.0 support is planned for the upcoming XLIFF 2.1 version.

A report summarizing the MultilingualWeb workshop in Madrid is now available from the MultilingualWeb site. It contains a summary of each session with links to presentation slides and minutes taken during the workshop in Madrid. The workshop was a huge success, with approximately 110 participants, and with the associated LIDER roadmapping workshop . The Workshop was hosted by Universidad Politécnica de Madrid , sponsored by the EU-funded LIDER project, by Verisign and by Lionbridge.
A new workshop in the MultilingualWeb series is planned for 2015.

This documentbuilds upon on the Character Model for the World Wide Web 1.0: Fundamentals to provide authors of specifications, software developers, and content developers a common reference on string matching on the World Wide Web and thereby increase interoperability. String matching is the process by which a specification or implementation defines whether two string values are the same or different from one another.

The main target audience of this specification is W3C specification developers. This specification and parts of it can be referenced from other W3C specifications and it defines conformance criteria for W3C specifications, as well as other specifications.

This version of this document represents a significant change from its previous edition. Much of the content is changed and the recommendations are significantly altered. This fact is reflected in a change to the name of the document from “Character Model: Normalization” to “Character Model for the World Wide Web: String Matching and Searching”.

The TAG will be holding a developer meetup along side of our next face-to-face meeting in Cambridge, Mass. The event will be in the evening of the 22nd of July, will be hosted by Akamai and organized by the BostonJS meetup group. Thanks to both Akamai and BostonJS for helping us out!

As with our previous TAG developer meet-ups, this will be a pretty simple format. We’ll get the TAG members up on stage for a panel discussion about some of the topics we’re covering. We’re going to let you know what we’re working on, answer questions and hopefully engage in some spirited discussion. The event is open to anyone interested in web architecture, web development, web standards and the future of web tech. It’s free to attend and you do not have to already be a member of BostonJS to attend (though you will have to join that meetup group if you want to register ). The event is also listed  on lanyrd (though you must register on the BostonJS meetup group to attend).

Version 7.0 of the Unicode Standardis now available, adding 2,834 new characters. This latest version adds the new currency symbols for the Russian ruble and Azerbaijani manat, approximately 250 emoji (pictographic symbols), many other symbols, and 23 new lesser-used and historic scripts, as well as character additions to many existing scripts. These additions extend support for written languages of North America, China, India, other Asian countries, and Africa. See the link above for full details.

Most of the new emoji characters derive from characters in long-standing and widespread use in Wingdings and Webdings fonts.

Major enhancements were made to the Indic script properties. New property values were added to enable a more algorithmic approach to rendering Indic scripts. These include properties for joining behavior, new classes for numbers, and a further division of the syllabic categories of viramas and rephas. With these enhancements, the default rendering for newly added Indic scripts can be significantly improved.

Unicode character properties were extended to the new characters. The old characters have enhancements to Script and Alphabetic properties, and casing and line-breaking behavior. There were also nearly 3,000 new Cantonese pronunciation entries, as well as new or clarified stability policies for promoting interoperable implementations.

Two other important Unicode specifications are maintained in synchrony with the Unicode Standard, and have updates for Version 7.0. These will be released at the same time:

UTS #10, Unicode Collation Algorithm— the standard for sorting Unicode text
UTS #46, Unicode IDNA Compatibility Processing— for processing of non-ASCII URLs (IDNs)

A Last Call Working Draft of Encodinghas been published.

While encodings have been defined to some extent, implementations have not always implemented them in the same way, have not always used the same labels, and often differ in dealing with undefined and former proprietary areas of encodings. This specification attempts to fill those gaps so that new implementations do not have to reverse engineer encoding implementations of the market leaders and existing implementations can converge.

The body of this spec is an exact copy of the WHATWG version as of the date of its publication, intended to provide a stable reference for other specifications. We are hoping for people to review the specification and send comments about any technical areas that need attention (see the Status section for details).

Please send comments by 1 July 2014.

The battle for web security and privacy is fought at many levels. Sometimes common practice in web application design can lead to data leakage with untended consequences for users. A good example of this came up recently where confidential files shared through common web-based document sharing services were being exposed unintentionaly to third parties because the private URLs used to share them had been unintentionally leaked.

URLs that allow a user to access an otherwise privileged resource or information are called  Capability URLs, and while they can be powerful, they can also cause potential problems when used improperly.

TAG member Jeni Tennison has been working on a draft defining the space of capability URLs and outlining some good practices for usage. We think this document should be useful for web builders who are thinking about incorporating this pattern into their applications. We think it’s pretty good, but we need your feedback before we finalize it and release it as a TAG finding.

The draft may be found here: http://www.w3.org/TR/capability-urls/ and if you have feedback you are encouraged to raise an issue on github or e-mail us on the TAG public mailing list. Thanks!

The Unicode Consortium is pleased to announce the release of version 2014-05-16 of the Unicode Ideographic Variation Database (IVD). This release registers the new Moji_Joho collection, along with the first 10,710 sequences in that collection, 9,685 of which are shared by the registered Hanyo-Denshi collection. Details can be found at http://www.unicode.org/ivd/.

The slides from the MultilingualWeb workshop (including several posters) and the LIDER roadmapping workshopare now available for download. Additional material (videos of the presentations, a workshop report and more) will follow in the next weeks – stay tuned.