From: Mark Davis โ (mark@macchiato.com)
Date: Sat Jan 30 2010 - 01:03:53 CST
Whoops... Thanks for the correction.
Mark
On Fri, Jan 29, 2010 at 21:30, Jungshik SHIN (์ ์ ์) <jshin1987@gmail.com>wrote:
> On Fri, Jan 29, 2010 at 11:06 AM, Mark Davis โ <mark@macchiato.com> wrote:
> > If anyone is interested, the detection algorithm we use is
> > open-sourced, available in the Chrome source code.
>
> Oops. Mark got confused. It's not yet open-sourced, but its 'cousin'
> for language detection was open-sourced as a part of Chrome. The
> language detector in Chrome is not yet in a shape to be separately
> used by other projects, though. We're planning to separate it out from
> the Chrome tree.
>
> Jungshik
>
>
> >
> > Mark
> >
> >
> >
> > On Fri, Jan 29, 2010 at 07:06, Mark Davis โ <mark@macchiato.com> wrote:
> >> It is encodings determined by a detection algorithm. The declarations
> >> for encodings (and language) are far too unreliable to be depended on.
> >> The detection algorithm itself is fairly complex, but quite fast and
> >> compact.
> >>
> >> Mark
> >>
> >>
> >>
> >> On Thu, Jan 28, 2010 at 21:38, Simon Montagu <smontagu@smontagu.org>
> wrote:
> >>> On 28/01/2010 10:50, Mark Davis โ wrote:
> >>>>
> >>>> There's a blog on Unicode that people may find interesting:
> >>>> http://googleblog.blogspot.com/2010/01/unicode-nearing-50-of-web.html
> >>>>
> >>>> (The graph on Unicode is too small; until they get that fixed, I have
> >>>> the large one on http://www.macchiato.com/)
> >>>>
> >>>> Mark
> >>>
> >>> What exactly is this counting? Encodings declared internally in
> web-pages?
> >>> Encodings declared in HTTP headers? Encodings determined by
> auto-detection?
> >>> Some combination of the above?
> >>>
> >>> --
> >>> Simon Montagu
> >>> Mozilla internationalization
> >>> ืกืืืืื ืืื ืืืื
> >>>
> >>>
> >>
> >
> >
> >
>
This archive was generated by hypermail 2.1.5 : Sat Jan 30 2010 - 01:09:36 CST