Great explanation of the rel attributes that provide:
- information to the resource linked
- behaviors while navigating
How to log performance information client-side.
Using an object model helps while designing some web pages or content.
An object in a web project is something that has structure, instances, and purpose. Instances are all the specific occurrences of an object.
An instance of the object link can be "When to choose a progressive web app".
So, as we make our user experiences more object-oriented, we also make them more intuitive.” - Sophia Prater, ooux.com.
With this practice, 4 unintuitive objects are identified. The author use the example of a zoo website and they are:
- isolated objects: they are disconnected from related objects. The penguin animals are not related to the demos and habitat from the list of all the animals. The habitat is also included on the map view of the zoo, but there are no links to the Animal or the demo on this map view either.
- broken objects: ones that are not directly manipulatable. Demos are useful to show events to the user. As someone who would like to attend the penguin demo, I want to click on an individual demo to see more details, but it is not possible.
- masked objects: styled to look the same, but are actually different. On the DisneyPlus web view, there is no way to differentiate between a series or a movie. As you want to see a movie (one shot) instead of a series (multiple shots), you have to check on the detailed view in a panel.
- shapeshifter objects: opposite of masked ones, they are styled to look different, but are actually the same. Date formatting, links, and so on should look and behave the same across the website or the experience platform. They have to be consistent.
The internet works because
- there is a stack of protocols built to make things work. Each protocol solves one thing.
- all miraculously work together because these standards are open.
Tim takes the example of the network stack: Ethernet Packet, Internet Packet, TCP Packet, the port, and the email protocols.
Protocols and standards are everywhere. He takes more examples. When you publish a web page for example: it can be both human and machine-readable. It can be accessed through a URI and when someone follows a link to your web page, their browser opens a TCP/IP connection to TCP port 80 on the machine which is registered as serving the (www.whatever.com, etc) in question. All of that is because the URI specification says that what you can tell about a URI depends on the first bit, in this case HTTP. Tim explains in depth why these relationships exist.
An XML document is a less specified version of an HTML document. The namespace declaration gives a URI indicating the namespace used to interpret this XML though. And more...
An RDF document is based on XML and a triple: a value of some property of some object, or some relationship between some object and some other object. How to figure out what a triple means? A URI defines it, and its standard can be read while dereferencing it. The color example is great! So the stack for this document piles up from the Ethernet to the RDF MS 1.0 and RDF MS 1.0 definition of rdf:type.
A pattern is that each technology evolves into three stages: using numbers or strings, then using a URI, and then a dereferenceable URI. As we move on to later protocols, the protocols themselves become more diverse, so URIs were created instead of simple versions with numbers or strings. "The third stage of civilization is the one at which the identifiers can be looked up on the web.".
This stack prevents one from sending a nasty email to someone and then protesting that the message didn't mean anything. So if the stack is so strict, how does one send a nasty email message when one doesn't mean it? Many protocols have ways of breaking the chain, of including information that is not part of the meaning of the message.
For the email: an attachment. "So being able to refer to something without asserting it, whether you call it attachment, packaging, or quoting, is an important feature of a language. The fact that you can do this removes the last excuse for anyone claiming not to have meant whatever they did say in the main message!"
Past a URL and retrieves some information about it.
First: check if the url is "safe"
A mission to accomplish: Working with the guidelines established by the manifesto and the social etiquette to progressively transform the culture of the internet and beyond.
There are three commitments to it:
- The commitment to social responsibility and partisanship
- The commitment to collective well-being and personal growth
- The commitment to rehumanizing social relations and reversing the process of social alienation
Old school GIFs put on website of the 2000
The headlines of the content picture well what content is in this post:
- The web used to be weirder
- Animated GIFs (WordArt)
- Music
- Cursor trails
- Webrings
- Guestbooks
- The majority of your users don't have access needs
- Accessibility is optional
- Access needs come from permanent disabilities
- Accessibility is a barrier to good design
- Accessibility is hard to implement
- React apps are inherently inaccessible
7 Automated testing will catch all accessibility problems
To my surprise they didn’t use skip links when they were presented one. [...] They didn’t understand the purpose of these links.
He explained that when he clicks on a link, for instance to an interesting article about skip links, he expects the first thing he encounters to be the article itself.
Instead of a "Skip to content", a "Skip to navigation" could be better.
It's an initiative for knowing the people behind a website. It's a TXT file that contains information about the different people who have contributed to building the website.
A txt file is simple and fast to create and can be embedded easily with a <link type="text/plain" rel="author" href="./humans.txt">
It completes robots.txt or security.txt
This project looks interesting.
- No VDOM
- Rust syntax
- CSS scoping
- Routing & SSR & SSG
- Tauri support
- Htmx integration
A website that embraces Brutalist Web Design is raw in its focus on content, and prioritization of the website visitor.
A website is about giving visitors content to enjoy and ways to interact with you.
Guidelines:
- Content is readable on all reasonable screens and devices.
- Only hyperlinks and buttons respond to clicks.
- Hyperlinks are underlined and buttons look like buttons.
- The back button works as expected.
- View content by scrolling.
- Decoration when needed and no unrelated content.
- Performance is a feature.
The only limitation of brutalist web design is that it is designed for content and not interactions. So this design method does not fit dashboards of real-time data for example. Am I missing something here?
I appreciate the guidelines that can be reused on every usable and accessible website or applications.
A quickstart tip:
Start with left-aligned black text on a white background, and to apply styling only to solve a specific problem
and more tips:
- Understand the semantic meaning of HTML elements.
- Learn about typography.
- Try designing for a small screen by default.
- Learn from designers about their choices and why they made them.
- When in doubt, do what Tron does: fight for the users.
Le premier lien web bleu tel qu'on le connais provient de
April 12, 1993 – Mosaic Version 0.13