If "AI art is kinda bad" is part of the argument against AI art, it will only be used against us when AI art isn't that bad anymore.
kogasa
It's a different situation, as a dev I'd happily bet my life on this assumption.
Dropping support for that stuff means breaking 95% of the websites people currently use. It's a non-starter, it cannot ever happen, even if you think it would be for the best.
Math builds up so much context that it's hard to avoid the use of shorthand and reused names for things. Every math book and paper will start with definitions. So it's not really on you for not recognizing it here
🍕(--, B) : C -> Set denotes the contravariant hom functor, normally written Hom(--, B). In this case, C is a category, and B is a fixed object in that category. The -- can be replaced by either an object or morphism of C, and that defines a map from C to Set.
For any given object X in C, the hom-set Hom(X, C) is the set of morphisms X -> B in C. For a morphism f : X -> Y in C, the Set morphism Hom(f, B) : Hom(Y, B) -> Hom(X, B) is defined by sending each g : Y -> B to gf : X -> B. This is the mapping C -> Set defined by Hom(--, C), and it's a (contravariant) functor because it respects composition: if h : X -> Y and f : Y -> Z then fh : X -> Z and Hom(fh, C) = Hom(h, C)Hom(f, C) sends g : Z -> B to gfh : X -> B.
--
P^(n)(R) AKA RP^n is the n-dimensional real projective space.
--
The caveat "phi is a morphism" is probably just to clarify that we're talking about "all morphisms X -> Y [in a given category]" and not simply all functions or something.
--
For more context, the derived functor of Hom(--, B) is called the Ext functor, and the exactness of that sequence (if the typo were fixed) is the statement of the universal coefficient theorem (for cohomology): https://en.wikipedia.org/wiki/Universal_coefficient_theorem The solution to this problem is the "Example: mod 2 cohomology of the real projective space" on that page. It's (Z/2Z)[x] / <x^(n+1)> or 🍔[x]/<x^(n+1)>, i.e. the ring of polynomials of degree n or less with coefficients in 🍔 = Z/2Z, meaning coefficients of 0 or 1.
It's not nonsense, although there is a typo that makes it technically unsolvable. If you fix the typo, it's an example calculation in the wikipedia page on the universal coefficient theorem: https://en.m.wikipedia.org/wiki/Universal_coefficient_theorem
It's real projective space
"put the excess energy into batteries" is an idea, and is already pretty much what is done, but the large scale implementation still requires a lot of time, effort, and expense.
Unless the improvements include making them fly at the same speed and height as fighter jets I'm not seeing the endgame.
The standard .NET C# compiler and CLI run on and build for Windows, MacOS, and Linux. You can run your ASP.NET webapps in a Linux docker container, or write console apps and run them on Linux, it doesn't matter anymore. As a .NET dev I have literally no reason to ever touch Windows, unless I'm touching legacy code from before .NET Core or building a Windows-exclusive app using a Windows app framework.
The GPT architecture is well understood, the part that is hard to explain is the way information is encoded in the trained model's parameters. It's not magic, it's just a highly opaque encoding.