Buckets:

download
raw
27.6 kB
import{s as ae,n as oe,o as se}from"../chunks/scheduler.852ec091.js";import{S as re,i as de,g as r,s as n,r as m,A as me,h as d,f as e,c as a,j as g,u as l,x as Y,k as h,y as s,a as o,v as p,d as b,t as c,w as _}from"../chunks/index.28275fd3.js";import{D as f}from"../chunks/Docstring.ee6c313e.js";import{H as Z,E as le}from"../chunks/EditOnGithub.582011f0.js";function pe(Qt){let u,dt,st,mt,E,lt,N,Xt='<a href="https://hf.co/papers/2302.06675" rel="nofollow">Lion (Evolved Sign Momentum)</a> is a unique optimizer that uses the sign of the gradient to determine the update direction of the momentum. This makes Lion more memory-efficient and faster than <code>AdamW</code> which tracks and store the first and second-order moments.',pt,A,bt,y,j,Wt,z,I,Ft,tt,Yt="Base Lion optimizer.",ct,V,_t,v,q,Mt,P,W,Ht,et,Zt="8-bit Lion optimizer.",gt,F,ht,$,M,St,T,H,Bt,it,te="32-bit Lion optimizer.",ft,S,ut,L,B,Gt,k,G,Ot,nt,ee="Paged Lion optimizer.",yt,O,vt,w,R,Rt,C,U,Ut,at,ie="Paged 8-bit Lion optimizer.",$t,J,Lt,x,K,Jt,D,Q,Kt,ot,ne="Paged 32-bit Lion optimizer.",wt,X,xt,rt,zt;return E=new Z({props:{title:"Lion",local:"lion",headingTag:"h1"}}),A=new Z({props:{title:"Lion",local:"api-class ][ bitsandbytes.optim.Lion",headingTag:"h2"}}),j=new f({props:{name:"class bitsandbytes.optim.Lion",anchor:"bitsandbytes.optim.Lion",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"optim_bits",val:" = 32"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"},{name:"is_paged",val:" = False"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L8"}}),I=new f({props:{name:"__init__",anchor:"bitsandbytes.optim.Lion.__init__",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"optim_bits",val:" = 32"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"},{name:"is_paged",val:" = False"}],parametersDescription:[{anchor:"bitsandbytes.optim.Lion.__init__.params",description:`<strong>params</strong> (<code>torch.tensor</code>) &#x2014;
The input parameters to optimize.`,name:"params"},{anchor:"bitsandbytes.optim.Lion.__init__.lr",description:`<strong>lr</strong> (<code>float</code>, defaults to 1e-4) &#x2014;
The learning rate.`,name:"lr"},{anchor:"bitsandbytes.optim.Lion.__init__.betas",description:`<strong>betas</strong> (<code>tuple(float, float)</code>, defaults to (0.9, 0.999)) &#x2014;
The beta values are the decay rates of the first and second-order moment of the optimizer.`,name:"betas"},{anchor:"bitsandbytes.optim.Lion.__init__.weight_decay",description:`<strong>weight_decay</strong> (<code>float</code>, defaults to 0) &#x2014;
The weight decay value for the optimizer.`,name:"weight_decay"},{anchor:"bitsandbytes.optim.Lion.__init__.optim_bits",description:`<strong>optim_bits</strong> (<code>int</code>, defaults to 32) &#x2014;
The number of bits of the optimizer state.`,name:"optim_bits"},{anchor:"bitsandbytes.optim.Lion.__init__.args",description:`<strong>args</strong> (<code>object</code>, defaults to <code>None</code>) &#x2014;
An object with additional arguments.`,name:"args"},{anchor:"bitsandbytes.optim.Lion.__init__.min_8bit_size",description:`<strong>min_8bit_size</strong> (<code>int</code>, defaults to 4096) &#x2014;
The minimum number of elements of the parameter tensors for 8-bit optimization.`,name:"min_8bit_size"},{anchor:"bitsandbytes.optim.Lion.__init__.percentile_clipping",description:`<strong>percentile_clipping</strong> (<code>int</code>, defaults to 100) &#x2014;
Adapts clipping threshold automatically by tracking the last 100 gradient norms and clipping the gradient at a certain percentile to improve stability.`,name:"percentile_clipping"},{anchor:"bitsandbytes.optim.Lion.__init__.block_wise",description:`<strong>block_wise</strong> (<code>bool</code>, defaults to <code>True</code>) &#x2014;
Whether to independently quantize each block of tensors to reduce outlier effects and improve stability.`,name:"block_wise"},{anchor:"bitsandbytes.optim.Lion.__init__.is_paged",description:`<strong>is_paged</strong> (<code>bool</code>, defaults to <code>False</code>) &#x2014;
Whether the optimizer is a paged optimizer or not.`,name:"is_paged"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L9"}}),V=new Z({props:{title:"Lion8bit",local:"bitsandbytes.optim.Lion8bit",headingTag:"h2"}}),q=new f({props:{name:"class bitsandbytes.optim.Lion8bit",anchor:"bitsandbytes.optim.Lion8bit",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"},{name:"is_paged",val:" = False"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L63"}}),W=new f({props:{name:"__init__",anchor:"bitsandbytes.optim.Lion8bit.__init__",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"},{name:"is_paged",val:" = False"}],parametersDescription:[{anchor:"bitsandbytes.optim.Lion8bit.__init__.params",description:`<strong>params</strong> (<code>torch.tensor</code>) &#x2014;
The input parameters to optimize.`,name:"params"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.lr",description:`<strong>lr</strong> (<code>float</code>, defaults to 1e-4) &#x2014;
The learning rate.`,name:"lr"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.betas",description:`<strong>betas</strong> (<code>tuple(float, float)</code>, defaults to (0.9, 0.999)) &#x2014;
The beta values are the decay rates of the first and second-order moment of the optimizer.`,name:"betas"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.weight_decay",description:`<strong>weight_decay</strong> (<code>float</code>, defaults to 0) &#x2014;
The weight decay value for the optimizer.`,name:"weight_decay"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.args",description:`<strong>args</strong> (<code>object</code>, defaults to <code>None</code>) &#x2014;
An object with additional arguments.`,name:"args"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.min_8bit_size",description:`<strong>min_8bit_size</strong> (<code>int</code>, defaults to 4096) &#x2014;
The minimum number of elements of the parameter tensors for 8-bit optimization.`,name:"min_8bit_size"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.percentile_clipping",description:`<strong>percentile_clipping</strong> (<code>int</code>, defaults to 100) &#x2014;
Adapts clipping threshold automatically by tracking the last 100 gradient norms and clipping the gradient at a certain percentile to improve stability.`,name:"percentile_clipping"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.block_wise",description:`<strong>block_wise</strong> (<code>bool</code>, defaults to <code>True</code>) &#x2014;
Whether to independently quantize each block of tensors to reduce outlier effects and improve stability.`,name:"block_wise"},{anchor:"bitsandbytes.optim.Lion8bit.__init__.is_paged",description:`<strong>is_paged</strong> (<code>bool</code>, defaults to <code>False</code>) &#x2014;
Whether the optimizer is a paged optimizer or not.`,name:"is_paged"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L64"}}),F=new Z({props:{title:"Lion32bit",local:"bitsandbytes.optim.Lion32bit",headingTag:"h2"}}),M=new f({props:{name:"class bitsandbytes.optim.Lion32bit",anchor:"bitsandbytes.optim.Lion32bit",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"},{name:"is_paged",val:" = False"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L115"}}),H=new f({props:{name:"__init__",anchor:"bitsandbytes.optim.Lion32bit.__init__",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"},{name:"is_paged",val:" = False"}],parametersDescription:[{anchor:"bitsandbytes.optim.Lion32bit.__init__.params",description:`<strong>params</strong> (<code>torch.tensor</code>) &#x2014;
The input parameters to optimize.`,name:"params"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.lr",description:`<strong>lr</strong> (<code>float</code>, defaults to 1e-4) &#x2014;
The learning rate.`,name:"lr"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.betas",description:`<strong>betas</strong> (<code>tuple(float, float)</code>, defaults to (0.9, 0.999)) &#x2014;
The beta values are the decay rates of the first and second-order moment of the optimizer.`,name:"betas"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.weight_decay",description:`<strong>weight_decay</strong> (<code>float</code>, defaults to 0) &#x2014;
The weight decay value for the optimizer.`,name:"weight_decay"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.args",description:`<strong>args</strong> (<code>object</code>, defaults to <code>None</code>) &#x2014;
An object with additional arguments.`,name:"args"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.min_8bit_size",description:`<strong>min_8bit_size</strong> (<code>int</code>, defaults to 4096) &#x2014;
The minimum number of elements of the parameter tensors for 8-bit optimization.`,name:"min_8bit_size"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.percentile_clipping",description:`<strong>percentile_clipping</strong> (<code>int</code>, defaults to 100) &#x2014;
Adapts clipping threshold automatically by tracking the last 100 gradient norms and clipping the gradient at a certain percentile to improve stability.`,name:"percentile_clipping"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.block_wise",description:`<strong>block_wise</strong> (<code>bool</code>, defaults to <code>True</code>) &#x2014;
Whether to independently quantize each block of tensors to reduce outlier effects and improve stability.`,name:"block_wise"},{anchor:"bitsandbytes.optim.Lion32bit.__init__.is_paged",description:`<strong>is_paged</strong> (<code>bool</code>, defaults to <code>False</code>) &#x2014;
Whether the optimizer is a paged optimizer or not.`,name:"is_paged"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L116"}}),S=new Z({props:{title:"PagedLion",local:"bitsandbytes.optim.PagedLion",headingTag:"h2"}}),B=new f({props:{name:"class bitsandbytes.optim.PagedLion",anchor:"bitsandbytes.optim.PagedLion",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"optim_bits",val:" = 32"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L167"}}),G=new f({props:{name:"__init__",anchor:"bitsandbytes.optim.PagedLion.__init__",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"optim_bits",val:" = 32"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"}],parametersDescription:[{anchor:"bitsandbytes.optim.PagedLion.__init__.params",description:`<strong>params</strong> (<code>torch.tensor</code>) &#x2014;
The input parameters to optimize.`,name:"params"},{anchor:"bitsandbytes.optim.PagedLion.__init__.lr",description:`<strong>lr</strong> (<code>float</code>, defaults to 1e-4) &#x2014;
The learning rate.`,name:"lr"},{anchor:"bitsandbytes.optim.PagedLion.__init__.betas",description:`<strong>betas</strong> (<code>tuple(float, float)</code>, defaults to (0.9, 0.999)) &#x2014;
The beta values are the decay rates of the first and second-order moment of the optimizer.`,name:"betas"},{anchor:"bitsandbytes.optim.PagedLion.__init__.weight_decay",description:`<strong>weight_decay</strong> (<code>float</code>, defaults to 0) &#x2014;
The weight decay value for the optimizer.`,name:"weight_decay"},{anchor:"bitsandbytes.optim.PagedLion.__init__.optim_bits",description:`<strong>optim_bits</strong> (<code>int</code>, defaults to 32) &#x2014;
The number of bits of the optimizer state.`,name:"optim_bits"},{anchor:"bitsandbytes.optim.PagedLion.__init__.args",description:`<strong>args</strong> (<code>object</code>, defaults to <code>None</code>) &#x2014;
An object with additional arguments.`,name:"args"},{anchor:"bitsandbytes.optim.PagedLion.__init__.min_8bit_size",description:`<strong>min_8bit_size</strong> (<code>int</code>, defaults to 4096) &#x2014;
The minimum number of elements of the parameter tensors for 8-bit optimization.`,name:"min_8bit_size"},{anchor:"bitsandbytes.optim.PagedLion.__init__.percentile_clipping",description:`<strong>percentile_clipping</strong> (<code>int</code>, defaults to 100) &#x2014;
Adapts clipping threshold automatically by tracking the last 100 gradient norms and clipping the gradient at a certain percentile to improve stability.`,name:"percentile_clipping"},{anchor:"bitsandbytes.optim.PagedLion.__init__.block_wise",description:`<strong>block_wise</strong> (<code>bool</code>, defaults to <code>True</code>) &#x2014;
Whether to independently quantize each block of tensors to reduce outlier effects and improve stability.`,name:"block_wise"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L168"}}),O=new Z({props:{title:"PagedLion8bit",local:"bitsandbytes.optim.PagedLion8bit",headingTag:"h2"}}),R=new f({props:{name:"class bitsandbytes.optim.PagedLion8bit",anchor:"bitsandbytes.optim.PagedLion8bit",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L219"}}),U=new f({props:{name:"__init__",anchor:"bitsandbytes.optim.PagedLion8bit.__init__",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"}],parametersDescription:[{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.params",description:`<strong>params</strong> (<code>torch.tensor</code>) &#x2014;
The input parameters to optimize.`,name:"params"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.lr",description:`<strong>lr</strong> (<code>float</code>, defaults to 1e-4) &#x2014;
The learning rate.`,name:"lr"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.betas",description:`<strong>betas</strong> (<code>tuple(float, float)</code>, defaults to (0.9, 0.999)) &#x2014;
The beta values are the decay rates of the first and second-order moment of the optimizer.`,name:"betas"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.weight_decay",description:`<strong>weight_decay</strong> (<code>float</code>, defaults to 0) &#x2014;
The weight decay value for the optimizer.`,name:"weight_decay"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.optim_bits",description:`<strong>optim_bits</strong> (<code>int</code>, defaults to 32) &#x2014;
The number of bits of the optimizer state.`,name:"optim_bits"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.args",description:`<strong>args</strong> (<code>object</code>, defaults to <code>None</code>) &#x2014;
An object with additional arguments.`,name:"args"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.min_8bit_size",description:`<strong>min_8bit_size</strong> (<code>int</code>, defaults to 4096) &#x2014;
The minimum number of elements of the parameter tensors for 8-bit optimization.`,name:"min_8bit_size"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.percentile_clipping",description:`<strong>percentile_clipping</strong> (<code>int</code>, defaults to 100) &#x2014;
Adapts clipping threshold automatically by tracking the last 100 gradient norms and clipping the gradient at a certain percentile to improve stability.`,name:"percentile_clipping"},{anchor:"bitsandbytes.optim.PagedLion8bit.__init__.block_wise",description:`<strong>block_wise</strong> (<code>bool</code>, defaults to <code>True</code>) &#x2014;
Whether to independently quantize each block of tensors to reduce outlier effects and improve stability.`,name:"block_wise"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L220"}}),J=new Z({props:{title:"PagedLion32bit",local:"bitsandbytes.optim.PagedLion32bit",headingTag:"h2"}}),K=new f({props:{name:"class bitsandbytes.optim.PagedLion32bit",anchor:"bitsandbytes.optim.PagedLion32bit",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L270"}}),Q=new f({props:{name:"__init__",anchor:"bitsandbytes.optim.PagedLion32bit.__init__",parameters:[{name:"params",val:""},{name:"lr",val:" = 0.0001"},{name:"betas",val:" = (0.9, 0.99)"},{name:"weight_decay",val:" = 0"},{name:"args",val:" = None"},{name:"min_8bit_size",val:" = 4096"},{name:"percentile_clipping",val:" = 100"},{name:"block_wise",val:" = True"}],parametersDescription:[{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.params",description:`<strong>params</strong> (<code>torch.tensor</code>) &#x2014;
The input parameters to optimize.`,name:"params"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.lr",description:`<strong>lr</strong> (<code>float</code>, defaults to 1e-4) &#x2014;
The learning rate.`,name:"lr"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.betas",description:`<strong>betas</strong> (<code>tuple(float, float)</code>, defaults to (0.9, 0.999)) &#x2014;
The beta values are the decay rates of the first and second-order moment of the optimizer.`,name:"betas"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.weight_decay",description:`<strong>weight_decay</strong> (<code>float</code>, defaults to 0) &#x2014;
The weight decay value for the optimizer.`,name:"weight_decay"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.optim_bits",description:`<strong>optim_bits</strong> (<code>int</code>, defaults to 32) &#x2014;
The number of bits of the optimizer state.`,name:"optim_bits"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.args",description:`<strong>args</strong> (<code>object</code>, defaults to <code>None</code>) &#x2014;
An object with additional arguments.`,name:"args"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.min_8bit_size",description:`<strong>min_8bit_size</strong> (<code>int</code>, defaults to 4096) &#x2014;
The minimum number of elements of the parameter tensors for 8-bit optimization.`,name:"min_8bit_size"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.percentile_clipping",description:`<strong>percentile_clipping</strong> (<code>int</code>, defaults to 100) &#x2014;
Adapts clipping threshold automatically by tracking the last 100 gradient norms and clipping the gradient at a certain percentile to improve stability.`,name:"percentile_clipping"},{anchor:"bitsandbytes.optim.PagedLion32bit.__init__.block_wise",description:`<strong>block_wise</strong> (<code>bool</code>, defaults to <code>True</code>) &#x2014;
Whether to independently quantize each block of tensors to reduce outlier effects and improve stability.`,name:"block_wise"}],source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/vr_1431/bitsandbytes/optim/lion.py#L271"}}),X=new le({props:{source:"https://github.com/bitsandbytes-foundation/bitsandbytes/blob/main/docs/source/reference/optim/lion.mdx"}}),{c(){u=r("meta"),dt=n(),st=r("p"),mt=n(),m(E.$$.fragment),lt=n(),N=r("p"),N.innerHTML=Xt,pt=n(),m(A.$$.fragment),bt=n(),y=r("div"),m(j.$$.fragment),Wt=n(),z=r("div"),m(I.$$.fragment),Ft=n(),tt=r("p"),tt.textContent=Yt,ct=n(),m(V.$$.fragment),_t=n(),v=r("div"),m(q.$$.fragment),Mt=n(),P=r("div"),m(W.$$.fragment),Ht=n(),et=r("p"),et.textContent=Zt,gt=n(),m(F.$$.fragment),ht=n(),$=r("div"),m(M.$$.fragment),St=n(),T=r("div"),m(H.$$.fragment),Bt=n(),it=r("p"),it.textContent=te,ft=n(),m(S.$$.fragment),ut=n(),L=r("div"),m(B.$$.fragment),Gt=n(),k=r("div"),m(G.$$.fragment),Ot=n(),nt=r("p"),nt.textContent=ee,yt=n(),m(O.$$.fragment),vt=n(),w=r("div"),m(R.$$.fragment),Rt=n(),C=r("div"),m(U.$$.fragment),Ut=n(),at=r("p"),at.textContent=ie,$t=n(),m(J.$$.fragment),Lt=n(),x=r("div"),m(K.$$.fragment),Jt=n(),D=r("div"),m(Q.$$.fragment),Kt=n(),ot=r("p"),ot.textContent=ne,wt=n(),m(X.$$.fragment),xt=n(),rt=r("p"),this.h()},l(t){const i=me("svelte-u9bgzb",document.head);u=d(i,"META",{name:!0,content:!0}),i.forEach(e),dt=a(t),st=d(t,"P",{}),g(st).forEach(e),mt=a(t),l(E.$$.fragment,t),lt=a(t),N=d(t,"P",{"data-svelte-h":!0}),Y(N)!=="svelte-1qfxmo"&&(N.innerHTML=Xt),pt=a(t),l(A.$$.fragment,t),bt=a(t),y=d(t,"DIV",{class:!0});var Pt=g(y);l(j.$$.fragment,Pt),Wt=a(Pt),z=d(Pt,"DIV",{class:!0});var Tt=g(z);l(I.$$.fragment,Tt),Ft=a(Tt),tt=d(Tt,"P",{"data-svelte-h":!0}),Y(tt)!=="svelte-zbrxqg"&&(tt.textContent=Yt),Tt.forEach(e),Pt.forEach(e),ct=a(t),l(V.$$.fragment,t),_t=a(t),v=d(t,"DIV",{class:!0});var kt=g(v);l(q.$$.fragment,kt),Mt=a(kt),P=d(kt,"DIV",{class:!0});var Ct=g(P);l(W.$$.fragment,Ct),Ht=a(Ct),et=d(Ct,"P",{"data-svelte-h":!0}),Y(et)!=="svelte-1t8hs73"&&(et.textContent=Zt),Ct.forEach(e),kt.forEach(e),gt=a(t),l(F.$$.fragment,t),ht=a(t),$=d(t,"DIV",{class:!0});var Dt=g($);l(M.$$.fragment,Dt),St=a(Dt),T=d(Dt,"DIV",{class:!0});var Et=g(T);l(H.$$.fragment,Et),Bt=a(Et),it=d(Et,"P",{"data-svelte-h":!0}),Y(it)!=="svelte-xjkt4u"&&(it.textContent=te),Et.forEach(e),Dt.forEach(e),ft=a(t),l(S.$$.fragment,t),ut=a(t),L=d(t,"DIV",{class:!0});var Nt=g(L);l(B.$$.fragment,Nt),Gt=a(Nt),k=d(Nt,"DIV",{class:!0});var At=g(k);l(G.$$.fragment,At),Ot=a(At),nt=d(At,"P",{"data-svelte-h":!0}),Y(nt)!=="svelte-fg4764"&&(nt.textContent=ee),At.forEach(e),Nt.forEach(e),yt=a(t),l(O.$$.fragment,t),vt=a(t),w=d(t,"DIV",{class:!0});var jt=g(w);l(R.$$.fragment,jt),Rt=a(jt),C=d(jt,"DIV",{class:!0});var It=g(C);l(U.$$.fragment,It),Ut=a(It),at=d(It,"P",{"data-svelte-h":!0}),Y(at)!=="svelte-1fu6xri"&&(at.textContent=ie),It.forEach(e),jt.forEach(e),$t=a(t),l(J.$$.fragment,t),Lt=a(t),x=d(t,"DIV",{class:!0});var Vt=g(x);l(K.$$.fragment,Vt),Jt=a(Vt),D=d(Vt,"DIV",{class:!0});var qt=g(D);l(Q.$$.fragment,qt),Kt=a(qt),ot=d(qt,"P",{"data-svelte-h":!0}),Y(ot)!=="svelte-1pg1i0n"&&(ot.textContent=ne),qt.forEach(e),Vt.forEach(e),wt=a(t),l(X.$$.fragment,t),xt=a(t),rt=d(t,"P",{}),g(rt).forEach(e),this.h()},h(){h(u,"name","hf:doc:metadata"),h(u,"content",be),h(z,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(y,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(P,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(v,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(T,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h($,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(k,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(L,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(C,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(w,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(D,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8"),h(x,"class","docstring border-l-2 border-t-2 pl-4 pt-3.5 border-gray-100 rounded-tl-xl mb-6 mt-8")},m(t,i){s(document.head,u),o(t,dt,i),o(t,st,i),o(t,mt,i),p(E,t,i),o(t,lt,i),o(t,N,i),o(t,pt,i),p(A,t,i),o(t,bt,i),o(t,y,i),p(j,y,null),s(y,Wt),s(y,z),p(I,z,null),s(z,Ft),s(z,tt),o(t,ct,i),p(V,t,i),o(t,_t,i),o(t,v,i),p(q,v,null),s(v,Mt),s(v,P),p(W,P,null),s(P,Ht),s(P,et),o(t,gt,i),p(F,t,i),o(t,ht,i),o(t,$,i),p(M,$,null),s($,St),s($,T),p(H,T,null),s(T,Bt),s(T,it),o(t,ft,i),p(S,t,i),o(t,ut,i),o(t,L,i),p(B,L,null),s(L,Gt),s(L,k),p(G,k,null),s(k,Ot),s(k,nt),o(t,yt,i),p(O,t,i),o(t,vt,i),o(t,w,i),p(R,w,null),s(w,Rt),s(w,C),p(U,C,null),s(C,Ut),s(C,at),o(t,$t,i),p(J,t,i),o(t,Lt,i),o(t,x,i),p(K,x,null),s(x,Jt),s(x,D),p(Q,D,null),s(D,Kt),s(D,ot),o(t,wt,i),p(X,t,i),o(t,xt,i),o(t,rt,i),zt=!0},p:oe,i(t){zt||(b(E.$$.fragment,t),b(A.$$.fragment,t),b(j.$$.fragment,t),b(I.$$.fragment,t),b(V.$$.fragment,t),b(q.$$.fragment,t),b(W.$$.fragment,t),b(F.$$.fragment,t),b(M.$$.fragment,t),b(H.$$.fragment,t),b(S.$$.fragment,t),b(B.$$.fragment,t),b(G.$$.fragment,t),b(O.$$.fragment,t),b(R.$$.fragment,t),b(U.$$.fragment,t),b(J.$$.fragment,t),b(K.$$.fragment,t),b(Q.$$.fragment,t),b(X.$$.fragment,t),zt=!0)},o(t){c(E.$$.fragment,t),c(A.$$.fragment,t),c(j.$$.fragment,t),c(I.$$.fragment,t),c(V.$$.fragment,t),c(q.$$.fragment,t),c(W.$$.fragment,t),c(F.$$.fragment,t),c(M.$$.fragment,t),c(H.$$.fragment,t),c(S.$$.fragment,t),c(B.$$.fragment,t),c(G.$$.fragment,t),c(O.$$.fragment,t),c(R.$$.fragment,t),c(U.$$.fragment,t),c(J.$$.fragment,t),c(K.$$.fragment,t),c(Q.$$.fragment,t),c(X.$$.fragment,t),zt=!1},d(t){t&&(e(dt),e(st),e(mt),e(lt),e(N),e(pt),e(bt),e(y),e(ct),e(_t),e(v),e(gt),e(ht),e($),e(ft),e(ut),e(L),e(yt),e(vt),e(w),e($t),e(Lt),e(x),e(wt),e(xt),e(rt)),e(u),_(E,t),_(A,t),_(j),_(I),_(V,t),_(q),_(W),_(F,t),_(M),_(H),_(S,t),_(B),_(G),_(O,t),_(R),_(U),_(J,t),_(K),_(Q),_(X,t)}}}const be='{"title":"Lion","local":"lion","sections":[{"title":"Lion","local":"api-class ][ bitsandbytes.optim.Lion","sections":[],"depth":2},{"title":"Lion8bit","local":"bitsandbytes.optim.Lion8bit","sections":[],"depth":2},{"title":"Lion32bit","local":"bitsandbytes.optim.Lion32bit","sections":[],"depth":2},{"title":"PagedLion","local":"bitsandbytes.optim.PagedLion","sections":[],"depth":2},{"title":"PagedLion8bit","local":"bitsandbytes.optim.PagedLion8bit","sections":[],"depth":2},{"title":"PagedLion32bit","local":"bitsandbytes.optim.PagedLion32bit","sections":[],"depth":2}],"depth":1}';function ce(Qt){return se(()=>{new URLSearchParams(window.location.search).get("fw")}),[]}class ue extends re{constructor(u){super(),de(this,u,ce,pe,ae,{})}}export{ue as component};

Xet Storage Details

Size:
27.6 kB
·
Xet hash:
e97f9d0ad0a497965891618aea0893fe9a3720d3a3c4e37576a4b1159ed320f6

Xet efficiently stores files, intelligently splitting them into unique chunks and accelerating uploads and downloads. More info.