From 78ec6779f465181103f329e4d81b74bb2bbe41f6 Mon Sep 17 00:00:00 2001 From: ltalirz Date: Sun, 24 Nov 2024 21:46:05 +0000 Subject: [PATCH] deploy: c5e0eeda07c7da293c0e9e7d7b2001ed61575747 --- assets/index-83daa10a.js | 1780 ++++++++++++++++++++++++++++++++++++++ index.html | 2 +- plugins_metadata.json | 34 +- 3 files changed, 1798 insertions(+), 18 deletions(-) create mode 100644 assets/index-83daa10a.js diff --git a/assets/index-83daa10a.js b/assets/index-83daa10a.js new file mode 100644 index 00000000..3e73daee --- /dev/null +++ b/assets/index-83daa10a.js @@ -0,0 +1,1780 @@ +function t0(e,t){for(var a=0;ai[s]})}}}return Object.freeze(Object.defineProperty(e,Symbol.toStringTag,{value:"Module"}))}(function(){const t=document.createElement("link").relList;if(t&&t.supports&&t.supports("modulepreload"))return;for(const s of document.querySelectorAll('link[rel="modulepreload"]'))i(s);new MutationObserver(s=>{for(const r of s)if(r.type==="childList")for(const n of r.addedNodes)n.tagName==="LINK"&&n.rel==="modulepreload"&&i(n)}).observe(document,{childList:!0,subtree:!0});function a(s){const r={};return s.integrity&&(r.integrity=s.integrity),s.referrerPolicy&&(r.referrerPolicy=s.referrerPolicy),s.crossOrigin==="use-credentials"?r.credentials="include":s.crossOrigin==="anonymous"?r.credentials="omit":r.credentials="same-origin",r}function i(s){if(s.ep)return;s.ep=!0;const r=a(s);fetch(s.href,r)}})();var ft=typeof globalThis<"u"?globalThis:typeof window<"u"?window:typeof global<"u"?global:typeof self<"u"?self:{};function gc(e){return e&&e.__esModule&&Object.prototype.hasOwnProperty.call(e,"default")?e.default:e}function Fa(e){if(e.__esModule)return e;var t=e.default;if(typeof t=="function"){var a=function i(){return this instanceof i?Reflect.construct(t,arguments,this.constructor):t.apply(this,arguments)};a.prototype=t.prototype}else a={};return Object.defineProperty(a,"__esModule",{value:!0}),Object.keys(e).forEach(function(i){var s=Object.getOwnPropertyDescriptor(e,i);Object.defineProperty(a,i,s.get?s:{enumerable:!0,get:function(){return e[i]}})}),a}var lg={exports:{}},Wo={},dg={exports:{}},Y={};/** + * @license React + * react.production.min.js + * + * Copyright (c) Facebook, Inc. and its affiliates. + * + * This source code is licensed under the MIT license found in the + * LICENSE file in the root directory of this source tree. + */var Hr=Symbol.for("react.element"),a0=Symbol.for("react.portal"),i0=Symbol.for("react.fragment"),s0=Symbol.for("react.strict_mode"),r0=Symbol.for("react.profiler"),n0=Symbol.for("react.provider"),o0=Symbol.for("react.context"),l0=Symbol.for("react.forward_ref"),d0=Symbol.for("react.suspense"),u0=Symbol.for("react.memo"),c0=Symbol.for("react.lazy"),cf=Symbol.iterator;function p0(e){return e===null||typeof e!="object"?null:(e=cf&&e[cf]||e["@@iterator"],typeof e=="function"?e:null)}var ug={isMounted:function(){return!1},enqueueForceUpdate:function(){},enqueueReplaceState:function(){},enqueueSetState:function(){}},cg=Object.assign,pg={};function Is(e,t,a){this.props=e,this.context=t,this.refs=pg,this.updater=a||ug}Is.prototype.isReactComponent={};Is.prototype.setState=function(e,t){if(typeof e!="object"&&typeof e!="function"&&e!=null)throw Error("setState(...): takes an object of state variables to update or a function which returns an object of state variables.");this.updater.enqueueSetState(this,e,t,"setState")};Is.prototype.forceUpdate=function(e){this.updater.enqueueForceUpdate(this,e,"forceUpdate")};function fg(){}fg.prototype=Is.prototype;function yc(e,t,a){this.props=e,this.context=t,this.refs=pg,this.updater=a||ug}var _c=yc.prototype=new fg;_c.constructor=yc;cg(_c,Is.prototype);_c.isPureReactComponent=!0;var pf=Array.isArray,mg=Object.prototype.hasOwnProperty,vc={current:null},hg={key:!0,ref:!0,__self:!0,__source:!0};function gg(e,t,a){var i,s={},r=null,n=null;if(t!=null)for(i in t.ref!==void 0&&(n=t.ref),t.key!==void 0&&(r=""+t.key),t)mg.call(t,i)&&!hg.hasOwnProperty(i)&&(s[i]=t[i]);var o=arguments.length-2;if(o===1)s.children=a;else if(1>>1,J=C[te];if(0>>1;tes(be,L))oes(He,be)?(C[te]=He,C[oe]=L,te=oe):(C[te]=be,C[ee]=L,te=ee);else if(oes(He,L))C[te]=He,C[oe]=L,te=oe;else break e}}return N}function s(C,N){var L=C.sortIndex-N.sortIndex;return L!==0?L:C.id-N.id}if(typeof performance=="object"&&typeof performance.now=="function"){var r=performance;e.unstable_now=function(){return r.now()}}else{var n=Date,o=n.now();e.unstable_now=function(){return n.now()-o}}var l=[],d=[],c=1,p=null,u=3,v=!1,f=!1,h=!1,T=typeof setTimeout=="function"?setTimeout:null,y=typeof clearTimeout=="function"?clearTimeout:null,g=typeof setImmediate<"u"?setImmediate:null;typeof navigator<"u"&&navigator.scheduling!==void 0&&navigator.scheduling.isInputPending!==void 0&&navigator.scheduling.isInputPending.bind(navigator.scheduling);function _(C){for(var N=a(d);N!==null;){if(N.callback===null)i(d);else if(N.startTime<=C)i(d),N.sortIndex=N.expirationTime,t(l,N);else break;N=a(d)}}function w(C){if(h=!1,_(C),!f)if(a(l)!==null)f=!0,O(D);else{var N=a(d);N!==null&&W(w,N.startTime-C)}}function D(C,N){f=!1,h&&(h=!1,y(S),S=-1),v=!0;var L=u;try{for(_(N),p=a(l);p!==null&&(!(p.expirationTime>N)||C&&!j());){var te=p.callback;if(typeof te=="function"){p.callback=null,u=p.priorityLevel;var J=te(p.expirationTime<=N);N=e.unstable_now(),typeof J=="function"?p.callback=J:p===a(l)&&i(l),_(N)}else i(l);p=a(l)}if(p!==null)var Ne=!0;else{var ee=a(d);ee!==null&&W(w,ee.startTime-N),Ne=!1}return Ne}finally{p=null,u=L,v=!1}}var x=!1,k=null,S=-1,A=5,P=-1;function j(){return!(e.unstable_now()-PC||125te?(C.sortIndex=L,t(d,C),a(l)===null&&C===a(d)&&(h?(y(S),S=-1):h=!0,W(w,L-te))):(C.sortIndex=J,t(l,C),f||v||(f=!0,O(D))),C},e.unstable_shouldYield=j,e.unstable_wrapCallback=function(C){var N=u;return function(){var L=u;u=N;try{return C.apply(this,arguments)}finally{u=L}}}})(wg);bg.exports=wg;var k0=bg.exports;/** + * @license React + * react-dom.production.min.js + * + * Copyright (c) Facebook, Inc. and its affiliates. + * + * This source code is licensed under the MIT license found in the + * LICENSE file in the root directory of this source tree. + */var q0=b,St=k0;function R(e){for(var t="https://reactjs.org/docs/error-decoder.html?invariant="+e,a=1;a"u"||typeof window.document>"u"||typeof window.document.createElement>"u"),Hd=Object.prototype.hasOwnProperty,D0=/^[:A-Z_a-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD][:A-Z_a-z\u00C0-\u00D6\u00D8-\u00F6\u00F8-\u02FF\u0370-\u037D\u037F-\u1FFF\u200C-\u200D\u2070-\u218F\u2C00-\u2FEF\u3001-\uD7FF\uF900-\uFDCF\uFDF0-\uFFFD\-.0-9\u00B7\u0300-\u036F\u203F-\u2040]*$/,mf={},hf={};function x0(e){return Hd.call(hf,e)?!0:Hd.call(mf,e)?!1:D0.test(e)?hf[e]=!0:(mf[e]=!0,!1)}function C0(e,t,a,i){if(a!==null&&a.type===0)return!1;switch(typeof t){case"function":case"symbol":return!0;case"boolean":return i?!1:a!==null?!a.acceptsBooleans:(e=e.toLowerCase().slice(0,5),e!=="data-"&&e!=="aria-");default:return!1}}function S0(e,t,a,i){if(t===null||typeof t>"u"||C0(e,t,a,i))return!0;if(i)return!1;if(a!==null)switch(a.type){case 3:return!t;case 4:return t===!1;case 5:return isNaN(t);case 6:return isNaN(t)||1>t}return!1}function lt(e,t,a,i,s,r,n){this.acceptsBooleans=t===2||t===3||t===4,this.attributeName=i,this.attributeNamespace=s,this.mustUseProperty=a,this.propertyName=e,this.type=t,this.sanitizeURL=r,this.removeEmptyString=n}var Ye={};"children dangerouslySetInnerHTML defaultValue defaultChecked innerHTML suppressContentEditableWarning suppressHydrationWarning style".split(" ").forEach(function(e){Ye[e]=new lt(e,0,!1,e,null,!1,!1)});[["acceptCharset","accept-charset"],["className","class"],["htmlFor","for"],["httpEquiv","http-equiv"]].forEach(function(e){var t=e[0];Ye[t]=new lt(t,1,!1,e[1],null,!1,!1)});["contentEditable","draggable","spellCheck","value"].forEach(function(e){Ye[e]=new lt(e,2,!1,e.toLowerCase(),null,!1,!1)});["autoReverse","externalResourcesRequired","focusable","preserveAlpha"].forEach(function(e){Ye[e]=new lt(e,2,!1,e,null,!1,!1)});"allowFullScreen async autoFocus autoPlay controls default defer disabled disablePictureInPicture disableRemotePlayback formNoValidate hidden loop noModule noValidate open playsInline readOnly required reversed scoped seamless itemScope".split(" ").forEach(function(e){Ye[e]=new lt(e,3,!1,e.toLowerCase(),null,!1,!1)});["checked","multiple","muted","selected"].forEach(function(e){Ye[e]=new lt(e,3,!0,e,null,!1,!1)});["capture","download"].forEach(function(e){Ye[e]=new lt(e,4,!1,e,null,!1,!1)});["cols","rows","size","span"].forEach(function(e){Ye[e]=new lt(e,6,!1,e,null,!1,!1)});["rowSpan","start"].forEach(function(e){Ye[e]=new lt(e,5,!1,e.toLowerCase(),null,!1,!1)});var wc=/[\-:]([a-z])/g;function Tc(e){return e[1].toUpperCase()}"accent-height alignment-baseline arabic-form baseline-shift cap-height clip-path clip-rule color-interpolation color-interpolation-filters color-profile color-rendering dominant-baseline enable-background fill-opacity fill-rule flood-color flood-opacity font-family font-size font-size-adjust font-stretch font-style font-variant font-weight glyph-name glyph-orientation-horizontal glyph-orientation-vertical horiz-adv-x horiz-origin-x image-rendering letter-spacing lighting-color marker-end marker-mid marker-start overline-position overline-thickness paint-order panose-1 pointer-events rendering-intent shape-rendering stop-color stop-opacity strikethrough-position strikethrough-thickness stroke-dasharray stroke-dashoffset stroke-linecap stroke-linejoin stroke-miterlimit stroke-opacity stroke-width text-anchor text-decoration text-rendering underline-position underline-thickness unicode-bidi unicode-range units-per-em v-alphabetic v-hanging v-ideographic v-mathematical vector-effect vert-adv-y vert-origin-x vert-origin-y word-spacing writing-mode xmlns:xlink x-height".split(" ").forEach(function(e){var t=e.replace(wc,Tc);Ye[t]=new lt(t,1,!1,e,null,!1,!1)});"xlink:actuate xlink:arcrole xlink:role xlink:show xlink:title xlink:type".split(" ").forEach(function(e){var t=e.replace(wc,Tc);Ye[t]=new lt(t,1,!1,e,"http://www.w3.org/1999/xlink",!1,!1)});["xml:base","xml:lang","xml:space"].forEach(function(e){var t=e.replace(wc,Tc);Ye[t]=new lt(t,1,!1,e,"http://www.w3.org/XML/1998/namespace",!1,!1)});["tabIndex","crossOrigin"].forEach(function(e){Ye[e]=new lt(e,1,!1,e.toLowerCase(),null,!1,!1)});Ye.xlinkHref=new lt("xlinkHref",1,!1,"xlink:href","http://www.w3.org/1999/xlink",!0,!1);["src","href","action","formAction"].forEach(function(e){Ye[e]=new lt(e,1,!1,e.toLowerCase(),null,!0,!0)});function kc(e,t,a,i){var s=Ye.hasOwnProperty(t)?Ye[t]:null;(s!==null?s.type!==0:i||!(2o||s[n]!==r[o]){var l=` +`+s[n].replace(" at new "," at ");return e.displayName&&l.includes("")&&(l=l.replace("",e.displayName)),l}while(1<=n&&0<=o);break}}}finally{ad=!1,Error.prepareStackTrace=a}return(e=e?e.displayName||e.name:"")?sr(e):""}function P0(e){switch(e.tag){case 5:return sr(e.type);case 16:return sr("Lazy");case 13:return sr("Suspense");case 19:return sr("SuspenseList");case 0:case 2:case 15:return e=id(e.type,!1),e;case 11:return e=id(e.type.render,!1),e;case 1:return e=id(e.type,!0),e;default:return""}}function Qd(e){if(e==null)return null;if(typeof e=="function")return e.displayName||e.name||null;if(typeof e=="string")return e;switch(e){case es:return"Fragment";case Zi:return"Portal";case Gd:return"Profiler";case qc:return"StrictMode";case Xd:return"Suspense";case Jd:return"SuspenseList"}if(typeof e=="object")switch(e.$$typeof){case qg:return(e.displayName||"Context")+".Consumer";case kg:return(e._context.displayName||"Context")+".Provider";case Dc:var t=e.render;return e=e.displayName,e||(e=t.displayName||t.name||"",e=e!==""?"ForwardRef("+e+")":"ForwardRef"),e;case xc:return t=e.displayName||null,t!==null?t:Qd(e.type)||"Memo";case Ka:t=e._payload,e=e._init;try{return Qd(e(t))}catch{}}return null}function A0(e){var t=e.type;switch(e.tag){case 24:return"Cache";case 9:return(t.displayName||"Context")+".Consumer";case 10:return(t._context.displayName||"Context")+".Provider";case 18:return"DehydratedFragment";case 11:return e=t.render,e=e.displayName||e.name||"",t.displayName||(e!==""?"ForwardRef("+e+")":"ForwardRef");case 7:return"Fragment";case 5:return t;case 4:return"Portal";case 3:return"Root";case 6:return"Text";case 16:return Qd(t);case 8:return t===qc?"StrictMode":"Mode";case 22:return"Offscreen";case 12:return"Profiler";case 21:return"Scope";case 13:return"Suspense";case 19:return"SuspenseList";case 25:return"TracingMarker";case 1:case 0:case 17:case 2:case 14:case 15:if(typeof t=="function")return t.displayName||t.name||null;if(typeof t=="string")return t}return null}function li(e){switch(typeof e){case"boolean":case"number":case"string":case"undefined":return e;case"object":return e;default:return""}}function xg(e){var t=e.type;return(e=e.nodeName)&&e.toLowerCase()==="input"&&(t==="checkbox"||t==="radio")}function R0(e){var t=xg(e)?"checked":"value",a=Object.getOwnPropertyDescriptor(e.constructor.prototype,t),i=""+e[t];if(!e.hasOwnProperty(t)&&typeof a<"u"&&typeof a.get=="function"&&typeof a.set=="function"){var s=a.get,r=a.set;return Object.defineProperty(e,t,{configurable:!0,get:function(){return s.call(this)},set:function(n){i=""+n,r.call(this,n)}}),Object.defineProperty(e,t,{enumerable:a.enumerable}),{getValue:function(){return i},setValue:function(n){i=""+n},stopTracking:function(){e._valueTracker=null,delete e[t]}}}}function fn(e){e._valueTracker||(e._valueTracker=R0(e))}function Cg(e){if(!e)return!1;var t=e._valueTracker;if(!t)return!0;var a=t.getValue(),i="";return e&&(i=xg(e)?e.checked?"true":"false":e.value),e=i,e!==a?(t.setValue(e),!0):!1}function ro(e){if(e=e||(typeof document<"u"?document:void 0),typeof e>"u")return null;try{return e.activeElement||e.body}catch{return e.body}}function Yd(e,t){var a=t.checked;return Pe({},t,{defaultChecked:void 0,defaultValue:void 0,value:void 0,checked:a??e._wrapperState.initialChecked})}function yf(e,t){var a=t.defaultValue==null?"":t.defaultValue,i=t.checked!=null?t.checked:t.defaultChecked;a=li(t.value!=null?t.value:a),e._wrapperState={initialChecked:i,initialValue:a,controlled:t.type==="checkbox"||t.type==="radio"?t.checked!=null:t.value!=null}}function Sg(e,t){t=t.checked,t!=null&&kc(e,"checked",t,!1)}function Zd(e,t){Sg(e,t);var a=li(t.value),i=t.type;if(a!=null)i==="number"?(a===0&&e.value===""||e.value!=a)&&(e.value=""+a):e.value!==""+a&&(e.value=""+a);else if(i==="submit"||i==="reset"){e.removeAttribute("value");return}t.hasOwnProperty("value")?eu(e,t.type,a):t.hasOwnProperty("defaultValue")&&eu(e,t.type,li(t.defaultValue)),t.checked==null&&t.defaultChecked!=null&&(e.defaultChecked=!!t.defaultChecked)}function _f(e,t,a){if(t.hasOwnProperty("value")||t.hasOwnProperty("defaultValue")){var i=t.type;if(!(i!=="submit"&&i!=="reset"||t.value!==void 0&&t.value!==null))return;t=""+e._wrapperState.initialValue,a||t===e.value||(e.value=t),e.defaultValue=t}a=e.name,a!==""&&(e.name=""),e.defaultChecked=!!e._wrapperState.initialChecked,a!==""&&(e.name=a)}function eu(e,t,a){(t!=="number"||ro(e.ownerDocument)!==e)&&(a==null?e.defaultValue=""+e._wrapperState.initialValue:e.defaultValue!==""+a&&(e.defaultValue=""+a))}var rr=Array.isArray;function ps(e,t,a,i){if(e=e.options,t){t={};for(var s=0;s"+t.valueOf().toString()+"",t=mn.firstChild;e.firstChild;)e.removeChild(e.firstChild);for(;t.firstChild;)e.appendChild(t.firstChild)}});function qr(e,t){if(t){var a=e.firstChild;if(a&&a===e.lastChild&&a.nodeType===3){a.nodeValue=t;return}}e.textContent=t}var ur={animationIterationCount:!0,aspectRatio:!0,borderImageOutset:!0,borderImageSlice:!0,borderImageWidth:!0,boxFlex:!0,boxFlexGroup:!0,boxOrdinalGroup:!0,columnCount:!0,columns:!0,flex:!0,flexGrow:!0,flexPositive:!0,flexShrink:!0,flexNegative:!0,flexOrder:!0,gridArea:!0,gridRow:!0,gridRowEnd:!0,gridRowSpan:!0,gridRowStart:!0,gridColumn:!0,gridColumnEnd:!0,gridColumnSpan:!0,gridColumnStart:!0,fontWeight:!0,lineClamp:!0,lineHeight:!0,opacity:!0,order:!0,orphans:!0,tabSize:!0,widows:!0,zIndex:!0,zoom:!0,fillOpacity:!0,floodOpacity:!0,stopOpacity:!0,strokeDasharray:!0,strokeDashoffset:!0,strokeMiterlimit:!0,strokeOpacity:!0,strokeWidth:!0},I0=["Webkit","ms","Moz","O"];Object.keys(ur).forEach(function(e){I0.forEach(function(t){t=t+e.charAt(0).toUpperCase()+e.substring(1),ur[t]=ur[e]})});function Ig(e,t,a){return t==null||typeof t=="boolean"||t===""?"":a||typeof t!="number"||t===0||ur.hasOwnProperty(e)&&ur[e]?(""+t).trim():t+"px"}function Ng(e,t){e=e.style;for(var a in t)if(t.hasOwnProperty(a)){var i=a.indexOf("--")===0,s=Ig(a,t[a],i);a==="float"&&(a="cssFloat"),i?e.setProperty(a,s):e[a]=s}}var N0=Pe({menuitem:!0},{area:!0,base:!0,br:!0,col:!0,embed:!0,hr:!0,img:!0,input:!0,keygen:!0,link:!0,meta:!0,param:!0,source:!0,track:!0,wbr:!0});function iu(e,t){if(t){if(N0[e]&&(t.children!=null||t.dangerouslySetInnerHTML!=null))throw Error(R(137,e));if(t.dangerouslySetInnerHTML!=null){if(t.children!=null)throw Error(R(60));if(typeof t.dangerouslySetInnerHTML!="object"||!("__html"in t.dangerouslySetInnerHTML))throw Error(R(61))}if(t.style!=null&&typeof t.style!="object")throw Error(R(62))}}function su(e,t){if(e.indexOf("-")===-1)return typeof t.is=="string";switch(e){case"annotation-xml":case"color-profile":case"font-face":case"font-face-src":case"font-face-uri":case"font-face-format":case"font-face-name":case"missing-glyph":return!1;default:return!0}}var ru=null;function Cc(e){return e=e.target||e.srcElement||window,e.correspondingUseElement&&(e=e.correspondingUseElement),e.nodeType===3?e.parentNode:e}var nu=null,fs=null,ms=null;function wf(e){if(e=Jr(e)){if(typeof nu!="function")throw Error(R(280));var t=e.stateNode;t&&(t=Vo(t),nu(e.stateNode,e.type,t))}}function jg(e){fs?ms?ms.push(e):ms=[e]:fs=e}function Eg(){if(fs){var e=fs,t=ms;if(ms=fs=null,wf(e),t)for(e=0;e>>=0,e===0?32:31-(U0(e)/K0|0)|0}var hn=64,gn=4194304;function nr(e){switch(e&-e){case 1:return 1;case 2:return 2;case 4:return 4;case 8:return 8;case 16:return 16;case 32:return 32;case 64:case 128:case 256:case 512:case 1024:case 2048:case 4096:case 8192:case 16384:case 32768:case 65536:case 131072:case 262144:case 524288:case 1048576:case 2097152:return e&4194240;case 4194304:case 8388608:case 16777216:case 33554432:case 67108864:return e&130023424;case 134217728:return 134217728;case 268435456:return 268435456;case 536870912:return 536870912;case 1073741824:return 1073741824;default:return e}}function uo(e,t){var a=e.pendingLanes;if(a===0)return 0;var i=0,s=e.suspendedLanes,r=e.pingedLanes,n=a&268435455;if(n!==0){var o=n&~s;o!==0?i=nr(o):(r&=n,r!==0&&(i=nr(r)))}else n=a&~s,n!==0?i=nr(n):r!==0&&(i=nr(r));if(i===0)return 0;if(t!==0&&t!==i&&!(t&s)&&(s=i&-i,r=t&-t,s>=r||s===16&&(r&4194240)!==0))return t;if(i&4&&(i|=a&16),t=e.entangledLanes,t!==0)for(e=e.entanglements,t&=i;0a;a++)t.push(e);return t}function Gr(e,t,a){e.pendingLanes|=t,t!==536870912&&(e.suspendedLanes=0,e.pingedLanes=0),e=e.eventTimes,t=31-ta(t),e[t]=a}function X0(e,t){var a=e.pendingLanes&~t;e.pendingLanes=t,e.suspendedLanes=0,e.pingedLanes=0,e.expiredLanes&=t,e.mutableReadLanes&=t,e.entangledLanes&=t,t=e.entanglements;var i=e.eventTimes;for(e=e.expirationTimes;0=pr),Af=String.fromCharCode(32),Rf=!1;function ty(e,t){switch(e){case"keyup":return kw.indexOf(t.keyCode)!==-1;case"keydown":return t.keyCode!==229;case"keypress":case"mousedown":case"focusout":return!0;default:return!1}}function ay(e){return e=e.detail,typeof e=="object"&&"data"in e?e.data:null}var ts=!1;function Dw(e,t){switch(e){case"compositionend":return ay(t);case"keypress":return t.which!==32?null:(Rf=!0,Af);case"textInput":return e=t.data,e===Af&&Rf?null:e;default:return null}}function xw(e,t){if(ts)return e==="compositionend"||!Ec&&ty(e,t)?(e=Zg(),zn=Ic=Ja=null,ts=!1,e):null;switch(e){case"paste":return null;case"keypress":if(!(t.ctrlKey||t.altKey||t.metaKey)||t.ctrlKey&&t.altKey){if(t.char&&1=t)return{node:a,offset:t-e};e=i}e:{for(;a;){if(a.nextSibling){a=a.nextSibling;break e}a=a.parentNode}a=void 0}a=Ef(a)}}function ny(e,t){return e&&t?e===t?!0:e&&e.nodeType===3?!1:t&&t.nodeType===3?ny(e,t.parentNode):"contains"in e?e.contains(t):e.compareDocumentPosition?!!(e.compareDocumentPosition(t)&16):!1:!1}function oy(){for(var e=window,t=ro();t instanceof e.HTMLIFrameElement;){try{var a=typeof t.contentWindow.location.href=="string"}catch{a=!1}if(a)e=t.contentWindow;else break;t=ro(e.document)}return t}function Fc(e){var t=e&&e.nodeName&&e.nodeName.toLowerCase();return t&&(t==="input"&&(e.type==="text"||e.type==="search"||e.type==="tel"||e.type==="url"||e.type==="password")||t==="textarea"||e.contentEditable==="true")}function Ew(e){var t=oy(),a=e.focusedElem,i=e.selectionRange;if(t!==a&&a&&a.ownerDocument&&ny(a.ownerDocument.documentElement,a)){if(i!==null&&Fc(a)){if(t=i.start,e=i.end,e===void 0&&(e=t),"selectionStart"in a)a.selectionStart=t,a.selectionEnd=Math.min(e,a.value.length);else if(e=(t=a.ownerDocument||document)&&t.defaultView||window,e.getSelection){e=e.getSelection();var s=a.textContent.length,r=Math.min(i.start,s);i=i.end===void 0?r:Math.min(i.end,s),!e.extend&&r>i&&(s=i,i=r,r=s),s=Ff(a,r);var n=Ff(a,i);s&&n&&(e.rangeCount!==1||e.anchorNode!==s.node||e.anchorOffset!==s.offset||e.focusNode!==n.node||e.focusOffset!==n.offset)&&(t=t.createRange(),t.setStart(s.node,s.offset),e.removeAllRanges(),r>i?(e.addRange(t),e.extend(n.node,n.offset)):(t.setEnd(n.node,n.offset),e.addRange(t)))}}for(t=[],e=a;e=e.parentNode;)e.nodeType===1&&t.push({element:e,left:e.scrollLeft,top:e.scrollTop});for(typeof a.focus=="function"&&a.focus(),a=0;a=document.documentMode,as=null,pu=null,mr=null,fu=!1;function Bf(e,t,a){var i=a.window===a?a.document:a.nodeType===9?a:a.ownerDocument;fu||as==null||as!==ro(i)||(i=as,"selectionStart"in i&&Fc(i)?i={start:i.selectionStart,end:i.selectionEnd}:(i=(i.ownerDocument&&i.ownerDocument.defaultView||window).getSelection(),i={anchorNode:i.anchorNode,anchorOffset:i.anchorOffset,focusNode:i.focusNode,focusOffset:i.focusOffset}),mr&&Ar(mr,i)||(mr=i,i=fo(pu,"onSelect"),0rs||(e.current=vu[rs],vu[rs]=null,rs--)}function ge(e,t){rs++,vu[rs]=e.current,e.current=t}var di={},st=ci(di),ht=ci(!1),Ii=di;function ws(e,t){var a=e.type.contextTypes;if(!a)return di;var i=e.stateNode;if(i&&i.__reactInternalMemoizedUnmaskedChildContext===t)return i.__reactInternalMemoizedMaskedChildContext;var s={},r;for(r in a)s[r]=t[r];return i&&(e=e.stateNode,e.__reactInternalMemoizedUnmaskedChildContext=t,e.__reactInternalMemoizedMaskedChildContext=s),s}function gt(e){return e=e.childContextTypes,e!=null}function ho(){ve(ht),ve(st)}function Uf(e,t,a){if(st.current!==di)throw Error(R(168));ge(st,t),ge(ht,a)}function gy(e,t,a){var i=e.stateNode;if(t=t.childContextTypes,typeof i.getChildContext!="function")return a;i=i.getChildContext();for(var s in i)if(!(s in t))throw Error(R(108,A0(e)||"Unknown",s));return Pe({},a,i)}function go(e){return e=(e=e.stateNode)&&e.__reactInternalMemoizedMergedChildContext||di,Ii=st.current,ge(st,e),ge(ht,ht.current),!0}function Kf(e,t,a){var i=e.stateNode;if(!i)throw Error(R(169));a?(e=gy(e,t,Ii),i.__reactInternalMemoizedMergedChildContext=e,ve(ht),ve(st),ge(st,e)):ve(ht),ge(ht,a)}var ka=null,Ho=!1,yd=!1;function yy(e){ka===null?ka=[e]:ka.push(e)}function Hw(e){Ho=!0,yy(e)}function pi(){if(!yd&&ka!==null){yd=!0;var e=0,t=le;try{var a=ka;for(le=1;e>=n,s-=n,qa=1<<32-ta(t)+s|a<S?(A=k,k=null):A=k.sibling;var P=u(y,k,_[S],w);if(P===null){k===null&&(k=A);break}e&&k&&P.alternate===null&&t(y,k),g=r(P,g,S),x===null?D=P:x.sibling=P,x=P,k=A}if(S===_.length)return a(y,k),qe&&_i(y,S),D;if(k===null){for(;S<_.length;S++)k=p(y,_[S],w),k!==null&&(g=r(k,g,S),x===null?D=k:x.sibling=k,x=k);return qe&&_i(y,S),D}for(k=i(y,k);S<_.length;S++)A=v(k,y,S,_[S],w),A!==null&&(e&&A.alternate!==null&&k.delete(A.key===null?S:A.key),g=r(A,g,S),x===null?D=A:x.sibling=A,x=A);return e&&k.forEach(function(j){return t(y,j)}),qe&&_i(y,S),D}function h(y,g,_,w){var D=zs(_);if(typeof D!="function")throw Error(R(150));if(_=D.call(_),_==null)throw Error(R(151));for(var x=D=null,k=g,S=g=0,A=null,P=_.next();k!==null&&!P.done;S++,P=_.next()){k.index>S?(A=k,k=null):A=k.sibling;var j=u(y,k,P.value,w);if(j===null){k===null&&(k=A);break}e&&k&&j.alternate===null&&t(y,k),g=r(j,g,S),x===null?D=j:x.sibling=j,x=j,k=A}if(P.done)return a(y,k),qe&&_i(y,S),D;if(k===null){for(;!P.done;S++,P=_.next())P=p(y,P.value,w),P!==null&&(g=r(P,g,S),x===null?D=P:x.sibling=P,x=P);return qe&&_i(y,S),D}for(k=i(y,k);!P.done;S++,P=_.next())P=v(k,y,S,P.value,w),P!==null&&(e&&P.alternate!==null&&k.delete(P.key===null?S:P.key),g=r(P,g,S),x===null?D=P:x.sibling=P,x=P);return e&&k.forEach(function(E){return t(y,E)}),qe&&_i(y,S),D}function T(y,g,_,w){if(typeof _=="object"&&_!==null&&_.type===es&&_.key===null&&(_=_.props.children),typeof _=="object"&&_!==null){switch(_.$$typeof){case pn:e:{for(var D=_.key,x=g;x!==null;){if(x.key===D){if(D=_.type,D===es){if(x.tag===7){a(y,x.sibling),g=s(x,_.props.children),g.return=y,y=g;break e}}else if(x.elementType===D||typeof D=="object"&&D!==null&&D.$$typeof===Ka&&Gf(D)===x.type){a(y,x.sibling),g=s(x,_.props),g.ref=Hs(y,x,_),g.return=y,y=g;break e}a(y,x);break}else t(y,x);x=x.sibling}_.type===es?(g=Ri(_.props.children,y.mode,w,_.key),g.return=y,y=g):(w=Jn(_.type,_.key,_.props,null,y.mode,w),w.ref=Hs(y,g,_),w.return=y,y=w)}return n(y);case Zi:e:{for(x=_.key;g!==null;){if(g.key===x)if(g.tag===4&&g.stateNode.containerInfo===_.containerInfo&&g.stateNode.implementation===_.implementation){a(y,g.sibling),g=s(g,_.children||[]),g.return=y,y=g;break e}else{a(y,g);break}else t(y,g);g=g.sibling}g=Dd(_,y.mode,w),g.return=y,y=g}return n(y);case Ka:return x=_._init,T(y,g,x(_._payload),w)}if(rr(_))return f(y,g,_,w);if(zs(_))return h(y,g,_,w);kn(y,_)}return typeof _=="string"&&_!==""||typeof _=="number"?(_=""+_,g!==null&&g.tag===6?(a(y,g.sibling),g=s(g,_),g.return=y,y=g):(a(y,g),g=qd(_,y.mode,w),g.return=y,y=g),n(y)):a(y,g)}return T}var ks=wy(!0),Ty=wy(!1),vo=ci(null),bo=null,ls=null,Lc=null;function Wc(){Lc=ls=bo=null}function zc(e){var t=vo.current;ve(vo),e._currentValue=t}function Tu(e,t,a){for(;e!==null;){var i=e.alternate;if((e.childLanes&t)!==t?(e.childLanes|=t,i!==null&&(i.childLanes|=t)):i!==null&&(i.childLanes&t)!==t&&(i.childLanes|=t),e===a)break;e=e.return}}function gs(e,t){bo=e,Lc=ls=null,e=e.dependencies,e!==null&&e.firstContext!==null&&(e.lanes&t&&(mt=!0),e.firstContext=null)}function Ut(e){var t=e._currentValue;if(Lc!==e)if(e={context:e,memoizedValue:t,next:null},ls===null){if(bo===null)throw Error(R(308));ls=e,bo.dependencies={lanes:0,firstContext:e}}else ls=ls.next=e;return t}var xi=null;function $c(e){xi===null?xi=[e]:xi.push(e)}function ky(e,t,a,i){var s=t.interleaved;return s===null?(a.next=a,$c(t)):(a.next=s.next,s.next=a),t.interleaved=a,Aa(e,i)}function Aa(e,t){e.lanes|=t;var a=e.alternate;for(a!==null&&(a.lanes|=t),a=e,e=e.return;e!==null;)e.childLanes|=t,a=e.alternate,a!==null&&(a.childLanes|=t),a=e,e=e.return;return a.tag===3?a.stateNode:null}var Va=!1;function Uc(e){e.updateQueue={baseState:e.memoizedState,firstBaseUpdate:null,lastBaseUpdate:null,shared:{pending:null,interleaved:null,lanes:0},effects:null}}function qy(e,t){e=e.updateQueue,t.updateQueue===e&&(t.updateQueue={baseState:e.baseState,firstBaseUpdate:e.firstBaseUpdate,lastBaseUpdate:e.lastBaseUpdate,shared:e.shared,effects:e.effects})}function Ca(e,t){return{eventTime:e,lane:t,tag:0,payload:null,callback:null,next:null}}function ii(e,t,a){var i=e.updateQueue;if(i===null)return null;if(i=i.shared,ae&2){var s=i.pending;return s===null?t.next=t:(t.next=s.next,s.next=t),i.pending=t,Aa(e,a)}return s=i.interleaved,s===null?(t.next=t,$c(i)):(t.next=s.next,s.next=t),i.interleaved=t,Aa(e,a)}function Un(e,t,a){if(t=t.updateQueue,t!==null&&(t=t.shared,(a&4194240)!==0)){var i=t.lanes;i&=e.pendingLanes,a|=i,t.lanes=a,Pc(e,a)}}function Xf(e,t){var a=e.updateQueue,i=e.alternate;if(i!==null&&(i=i.updateQueue,a===i)){var s=null,r=null;if(a=a.firstBaseUpdate,a!==null){do{var n={eventTime:a.eventTime,lane:a.lane,tag:a.tag,payload:a.payload,callback:a.callback,next:null};r===null?s=r=n:r=r.next=n,a=a.next}while(a!==null);r===null?s=r=t:r=r.next=t}else s=r=t;a={baseState:i.baseState,firstBaseUpdate:s,lastBaseUpdate:r,shared:i.shared,effects:i.effects},e.updateQueue=a;return}e=a.lastBaseUpdate,e===null?a.firstBaseUpdate=t:e.next=t,a.lastBaseUpdate=t}function wo(e,t,a,i){var s=e.updateQueue;Va=!1;var r=s.firstBaseUpdate,n=s.lastBaseUpdate,o=s.shared.pending;if(o!==null){s.shared.pending=null;var l=o,d=l.next;l.next=null,n===null?r=d:n.next=d,n=l;var c=e.alternate;c!==null&&(c=c.updateQueue,o=c.lastBaseUpdate,o!==n&&(o===null?c.firstBaseUpdate=d:o.next=d,c.lastBaseUpdate=l))}if(r!==null){var p=s.baseState;n=0,c=d=l=null,o=r;do{var u=o.lane,v=o.eventTime;if((i&u)===u){c!==null&&(c=c.next={eventTime:v,lane:0,tag:o.tag,payload:o.payload,callback:o.callback,next:null});e:{var f=e,h=o;switch(u=t,v=a,h.tag){case 1:if(f=h.payload,typeof f=="function"){p=f.call(v,p,u);break e}p=f;break e;case 3:f.flags=f.flags&-65537|128;case 0:if(f=h.payload,u=typeof f=="function"?f.call(v,p,u):f,u==null)break e;p=Pe({},p,u);break e;case 2:Va=!0}}o.callback!==null&&o.lane!==0&&(e.flags|=64,u=s.effects,u===null?s.effects=[o]:u.push(o))}else v={eventTime:v,lane:u,tag:o.tag,payload:o.payload,callback:o.callback,next:null},c===null?(d=c=v,l=p):c=c.next=v,n|=u;if(o=o.next,o===null){if(o=s.shared.pending,o===null)break;u=o,o=u.next,u.next=null,s.lastBaseUpdate=u,s.shared.pending=null}}while(1);if(c===null&&(l=p),s.baseState=l,s.firstBaseUpdate=d,s.lastBaseUpdate=c,t=s.shared.interleaved,t!==null){s=t;do n|=s.lane,s=s.next;while(s!==t)}else r===null&&(s.shared.lanes=0);Ei|=n,e.lanes=n,e.memoizedState=p}}function Jf(e,t,a){if(e=t.effects,t.effects=null,e!==null)for(t=0;ta?a:4,e(!0);var i=vd.transition;vd.transition={};try{e(!1),t()}finally{le=a,vd.transition=i}}function Wy(){return Kt().memoizedState}function Qw(e,t,a){var i=ri(e);if(a={lane:i,action:a,hasEagerState:!1,eagerState:null,next:null},zy(e))$y(t,a);else if(a=ky(e,t,a,i),a!==null){var s=nt();aa(a,e,i,s),Uy(a,t,i)}}function Yw(e,t,a){var i=ri(e),s={lane:i,action:a,hasEagerState:!1,eagerState:null,next:null};if(zy(e))$y(t,s);else{var r=e.alternate;if(e.lanes===0&&(r===null||r.lanes===0)&&(r=t.lastRenderedReducer,r!==null))try{var n=t.lastRenderedState,o=r(n,a);if(s.hasEagerState=!0,s.eagerState=o,ia(o,n)){var l=t.interleaved;l===null?(s.next=s,$c(t)):(s.next=l.next,l.next=s),t.interleaved=s;return}}catch{}finally{}a=ky(e,t,s,i),a!==null&&(s=nt(),aa(a,e,i,s),Uy(a,t,i))}}function zy(e){var t=e.alternate;return e===Se||t!==null&&t===Se}function $y(e,t){hr=ko=!0;var a=e.pending;a===null?t.next=t:(t.next=a.next,a.next=t),e.pending=t}function Uy(e,t,a){if(a&4194240){var i=t.lanes;i&=e.pendingLanes,a|=i,t.lanes=a,Pc(e,a)}}var qo={readContext:Ut,useCallback:et,useContext:et,useEffect:et,useImperativeHandle:et,useInsertionEffect:et,useLayoutEffect:et,useMemo:et,useReducer:et,useRef:et,useState:et,useDebugValue:et,useDeferredValue:et,useTransition:et,useMutableSource:et,useSyncExternalStore:et,useId:et,unstable_isNewReconciler:!1},Zw={readContext:Ut,useCallback:function(e,t){return oa().memoizedState=[e,t===void 0?null:t],e},useContext:Ut,useEffect:Yf,useImperativeHandle:function(e,t,a){return a=a!=null?a.concat([e]):null,Vn(4194308,4,Fy.bind(null,t,e),a)},useLayoutEffect:function(e,t){return Vn(4194308,4,e,t)},useInsertionEffect:function(e,t){return Vn(4,2,e,t)},useMemo:function(e,t){var a=oa();return t=t===void 0?null:t,e=e(),a.memoizedState=[e,t],e},useReducer:function(e,t,a){var i=oa();return t=a!==void 0?a(t):t,i.memoizedState=i.baseState=t,e={pending:null,interleaved:null,lanes:0,dispatch:null,lastRenderedReducer:e,lastRenderedState:t},i.queue=e,e=e.dispatch=Qw.bind(null,Se,e),[i.memoizedState,e]},useRef:function(e){var t=oa();return e={current:e},t.memoizedState=e},useState:Qf,useDebugValue:Yc,useDeferredValue:function(e){return oa().memoizedState=e},useTransition:function(){var e=Qf(!1),t=e[0];return e=Jw.bind(null,e[1]),oa().memoizedState=e,[t,e]},useMutableSource:function(){},useSyncExternalStore:function(e,t,a){var i=Se,s=oa();if(qe){if(a===void 0)throw Error(R(407));a=a()}else{if(a=t(),Ue===null)throw Error(R(349));ji&30||Sy(i,t,a)}s.memoizedState=a;var r={value:a,getSnapshot:t};return s.queue=r,Yf(Ay.bind(null,i,r,e),[e]),i.flags|=2048,Or(9,Py.bind(null,i,r,a,t),void 0,null),a},useId:function(){var e=oa(),t=Ue.identifierPrefix;if(qe){var a=Da,i=qa;a=(i&~(1<<32-ta(i)-1)).toString(32)+a,t=":"+t+"R"+a,a=Fr++,0<\/script>",e=e.removeChild(e.firstChild)):typeof i.is=="string"?e=n.createElement(a,{is:i.is}):(e=n.createElement(a),a==="select"&&(n=e,i.multiple?n.multiple=!0:i.size&&(n.size=i.size))):e=n.createElementNS(e,a),e[ua]=t,e[Nr]=i,e_(e,t,!1,!1),t.stateNode=e;e:{switch(n=su(a,i),a){case"dialog":_e("cancel",e),_e("close",e),s=i;break;case"iframe":case"object":case"embed":_e("load",e),s=i;break;case"video":case"audio":for(s=0;sxs&&(t.flags|=128,i=!0,Gs(r,!1),t.lanes=4194304)}else{if(!i)if(e=To(n),e!==null){if(t.flags|=128,i=!0,a=e.updateQueue,a!==null&&(t.updateQueue=a,t.flags|=4),Gs(r,!0),r.tail===null&&r.tailMode==="hidden"&&!n.alternate&&!qe)return tt(t),null}else 2*Ee()-r.renderingStartTime>xs&&a!==1073741824&&(t.flags|=128,i=!0,Gs(r,!1),t.lanes=4194304);r.isBackwards?(n.sibling=t.child,t.child=n):(a=r.last,a!==null?a.sibling=n:t.child=n,r.last=n)}return r.tail!==null?(t=r.tail,r.rendering=t,r.tail=t.sibling,r.renderingStartTime=Ee(),t.sibling=null,a=Ce.current,ge(Ce,i?a&1|2:a&1),t):(tt(t),null);case 22:case 23:return sp(),i=t.memoizedState!==null,e!==null&&e.memoizedState!==null!==i&&(t.flags|=8192),i&&t.mode&1?wt&1073741824&&(tt(t),t.subtreeFlags&6&&(t.flags|=8192)):tt(t),null;case 24:return null;case 25:return null}throw Error(R(156,t.tag))}function o1(e,t){switch(Oc(t),t.tag){case 1:return gt(t.type)&&ho(),e=t.flags,e&65536?(t.flags=e&-65537|128,t):null;case 3:return qs(),ve(ht),ve(st),Hc(),e=t.flags,e&65536&&!(e&128)?(t.flags=e&-65537|128,t):null;case 5:return Vc(t),null;case 13:if(ve(Ce),e=t.memoizedState,e!==null&&e.dehydrated!==null){if(t.alternate===null)throw Error(R(340));Ts()}return e=t.flags,e&65536?(t.flags=e&-65537|128,t):null;case 19:return ve(Ce),null;case 4:return qs(),null;case 10:return zc(t.type._context),null;case 22:case 23:return sp(),null;case 24:return null;default:return null}}var Dn=!1,it=!1,l1=typeof WeakSet=="function"?WeakSet:Set,M=null;function ds(e,t){var a=e.ref;if(a!==null)if(typeof a=="function")try{a(null)}catch(i){Ie(e,t,i)}else a.current=null}function Ru(e,t,a){try{a()}catch(i){Ie(e,t,i)}}var dm=!1;function d1(e,t){if(mu=co,e=oy(),Fc(e)){if("selectionStart"in e)var a={start:e.selectionStart,end:e.selectionEnd};else e:{a=(a=e.ownerDocument)&&a.defaultView||window;var i=a.getSelection&&a.getSelection();if(i&&i.rangeCount!==0){a=i.anchorNode;var s=i.anchorOffset,r=i.focusNode;i=i.focusOffset;try{a.nodeType,r.nodeType}catch{a=null;break e}var n=0,o=-1,l=-1,d=0,c=0,p=e,u=null;t:for(;;){for(var v;p!==a||s!==0&&p.nodeType!==3||(o=n+s),p!==r||i!==0&&p.nodeType!==3||(l=n+i),p.nodeType===3&&(n+=p.nodeValue.length),(v=p.firstChild)!==null;)u=p,p=v;for(;;){if(p===e)break t;if(u===a&&++d===s&&(o=n),u===r&&++c===i&&(l=n),(v=p.nextSibling)!==null)break;p=u,u=p.parentNode}p=v}a=o===-1||l===-1?null:{start:o,end:l}}else a=null}a=a||{start:0,end:0}}else a=null;for(hu={focusedElem:e,selectionRange:a},co=!1,M=t;M!==null;)if(t=M,e=t.child,(t.subtreeFlags&1028)!==0&&e!==null)e.return=t,M=e;else for(;M!==null;){t=M;try{var f=t.alternate;if(t.flags&1024)switch(t.tag){case 0:case 11:case 15:break;case 1:if(f!==null){var h=f.memoizedProps,T=f.memoizedState,y=t.stateNode,g=y.getSnapshotBeforeUpdate(t.elementType===t.type?h:Qt(t.type,h),T);y.__reactInternalSnapshotBeforeUpdate=g}break;case 3:var _=t.stateNode.containerInfo;_.nodeType===1?_.textContent="":_.nodeType===9&&_.documentElement&&_.removeChild(_.documentElement);break;case 5:case 6:case 4:case 17:break;default:throw Error(R(163))}}catch(w){Ie(t,t.return,w)}if(e=t.sibling,e!==null){e.return=t.return,M=e;break}M=t.return}return f=dm,dm=!1,f}function gr(e,t,a){var i=t.updateQueue;if(i=i!==null?i.lastEffect:null,i!==null){var s=i=i.next;do{if((s.tag&e)===e){var r=s.destroy;s.destroy=void 0,r!==void 0&&Ru(t,a,r)}s=s.next}while(s!==i)}}function Jo(e,t){if(t=t.updateQueue,t=t!==null?t.lastEffect:null,t!==null){var a=t=t.next;do{if((a.tag&e)===e){var i=a.create;a.destroy=i()}a=a.next}while(a!==t)}}function Iu(e){var t=e.ref;if(t!==null){var a=e.stateNode;switch(e.tag){case 5:e=a;break;default:e=a}typeof t=="function"?t(e):t.current=e}}function i_(e){var t=e.alternate;t!==null&&(e.alternate=null,i_(t)),e.child=null,e.deletions=null,e.sibling=null,e.tag===5&&(t=e.stateNode,t!==null&&(delete t[ua],delete t[Nr],delete t[_u],delete t[Kw],delete t[Vw])),e.stateNode=null,e.return=null,e.dependencies=null,e.memoizedProps=null,e.memoizedState=null,e.pendingProps=null,e.stateNode=null,e.updateQueue=null}function s_(e){return e.tag===5||e.tag===3||e.tag===4}function um(e){e:for(;;){for(;e.sibling===null;){if(e.return===null||s_(e.return))return null;e=e.return}for(e.sibling.return=e.return,e=e.sibling;e.tag!==5&&e.tag!==6&&e.tag!==18;){if(e.flags&2||e.child===null||e.tag===4)continue e;e.child.return=e,e=e.child}if(!(e.flags&2))return e.stateNode}}function Nu(e,t,a){var i=e.tag;if(i===5||i===6)e=e.stateNode,t?a.nodeType===8?a.parentNode.insertBefore(e,t):a.insertBefore(e,t):(a.nodeType===8?(t=a.parentNode,t.insertBefore(e,a)):(t=a,t.appendChild(e)),a=a._reactRootContainer,a!=null||t.onclick!==null||(t.onclick=mo));else if(i!==4&&(e=e.child,e!==null))for(Nu(e,t,a),e=e.sibling;e!==null;)Nu(e,t,a),e=e.sibling}function ju(e,t,a){var i=e.tag;if(i===5||i===6)e=e.stateNode,t?a.insertBefore(e,t):a.appendChild(e);else if(i!==4&&(e=e.child,e!==null))for(ju(e,t,a),e=e.sibling;e!==null;)ju(e,t,a),e=e.sibling}var Xe=null,Yt=!1;function Wa(e,t,a){for(a=a.child;a!==null;)r_(e,t,a),a=a.sibling}function r_(e,t,a){if(pa&&typeof pa.onCommitFiberUnmount=="function")try{pa.onCommitFiberUnmount(zo,a)}catch{}switch(a.tag){case 5:it||ds(a,t);case 6:var i=Xe,s=Yt;Xe=null,Wa(e,t,a),Xe=i,Yt=s,Xe!==null&&(Yt?(e=Xe,a=a.stateNode,e.nodeType===8?e.parentNode.removeChild(a):e.removeChild(a)):Xe.removeChild(a.stateNode));break;case 18:Xe!==null&&(Yt?(e=Xe,a=a.stateNode,e.nodeType===8?gd(e.parentNode,a):e.nodeType===1&&gd(e,a),Sr(e)):gd(Xe,a.stateNode));break;case 4:i=Xe,s=Yt,Xe=a.stateNode.containerInfo,Yt=!0,Wa(e,t,a),Xe=i,Yt=s;break;case 0:case 11:case 14:case 15:if(!it&&(i=a.updateQueue,i!==null&&(i=i.lastEffect,i!==null))){s=i=i.next;do{var r=s,n=r.destroy;r=r.tag,n!==void 0&&(r&2||r&4)&&Ru(a,t,n),s=s.next}while(s!==i)}Wa(e,t,a);break;case 1:if(!it&&(ds(a,t),i=a.stateNode,typeof i.componentWillUnmount=="function"))try{i.props=a.memoizedProps,i.state=a.memoizedState,i.componentWillUnmount()}catch(o){Ie(a,t,o)}Wa(e,t,a);break;case 21:Wa(e,t,a);break;case 22:a.mode&1?(it=(i=it)||a.memoizedState!==null,Wa(e,t,a),it=i):Wa(e,t,a);break;default:Wa(e,t,a)}}function cm(e){var t=e.updateQueue;if(t!==null){e.updateQueue=null;var a=e.stateNode;a===null&&(a=e.stateNode=new l1),t.forEach(function(i){var s=_1.bind(null,e,i);a.has(i)||(a.add(i),i.then(s,s))})}}function Jt(e,t){var a=t.deletions;if(a!==null)for(var i=0;is&&(s=n),i&=~r}if(i=s,i=Ee()-i,i=(120>i?120:480>i?480:1080>i?1080:1920>i?1920:3e3>i?3e3:4320>i?4320:1960*c1(i/1960))-i,10e?16:e,Qa===null)var i=!1;else{if(e=Qa,Qa=null,Co=0,ae&6)throw Error(R(331));var s=ae;for(ae|=4,M=e.current;M!==null;){var r=M,n=r.child;if(M.flags&16){var o=r.deletions;if(o!==null){for(var l=0;lEe()-ap?Ai(e,0):tp|=a),yt(e,t)}function f_(e,t){t===0&&(e.mode&1?(t=gn,gn<<=1,!(gn&130023424)&&(gn=4194304)):t=1);var a=nt();e=Aa(e,t),e!==null&&(Gr(e,t,a),yt(e,a))}function y1(e){var t=e.memoizedState,a=0;t!==null&&(a=t.retryLane),f_(e,a)}function _1(e,t){var a=0;switch(e.tag){case 13:var i=e.stateNode,s=e.memoizedState;s!==null&&(a=s.retryLane);break;case 19:i=e.stateNode;break;default:throw Error(R(314))}i!==null&&i.delete(t),f_(e,a)}var m_;m_=function(e,t,a){if(e!==null)if(e.memoizedProps!==t.pendingProps||ht.current)mt=!0;else{if(!(e.lanes&a)&&!(t.flags&128))return mt=!1,r1(e,t,a);mt=!!(e.flags&131072)}else mt=!1,qe&&t.flags&1048576&&_y(t,_o,t.index);switch(t.lanes=0,t.tag){case 2:var i=t.type;Hn(e,t),e=t.pendingProps;var s=ws(t,st.current);gs(t,a),s=Xc(null,t,i,e,s,a);var r=Jc();return t.flags|=1,typeof s=="object"&&s!==null&&typeof s.render=="function"&&s.$$typeof===void 0?(t.tag=1,t.memoizedState=null,t.updateQueue=null,gt(i)?(r=!0,go(t)):r=!1,t.memoizedState=s.state!==null&&s.state!==void 0?s.state:null,Uc(t),s.updater=Xo,t.stateNode=s,s._reactInternals=t,qu(t,i,e,a),t=Cu(null,t,i,!0,r,a)):(t.tag=0,qe&&r&&Bc(t),rt(null,t,s,a),t=t.child),t;case 16:i=t.elementType;e:{switch(Hn(e,t),e=t.pendingProps,s=i._init,i=s(i._payload),t.type=i,s=t.tag=b1(i),e=Qt(i,e),s){case 0:t=xu(null,t,i,e,a);break e;case 1:t=nm(null,t,i,e,a);break e;case 11:t=sm(null,t,i,e,a);break e;case 14:t=rm(null,t,i,Qt(i.type,e),a);break e}throw Error(R(306,i,""))}return t;case 0:return i=t.type,s=t.pendingProps,s=t.elementType===i?s:Qt(i,s),xu(e,t,i,s,a);case 1:return i=t.type,s=t.pendingProps,s=t.elementType===i?s:Qt(i,s),nm(e,t,i,s,a);case 3:e:{if(Qy(t),e===null)throw Error(R(387));i=t.pendingProps,r=t.memoizedState,s=r.element,qy(e,t),wo(t,i,null,a);var n=t.memoizedState;if(i=n.element,r.isDehydrated)if(r={element:i,isDehydrated:!1,cache:n.cache,pendingSuspenseBoundaries:n.pendingSuspenseBoundaries,transitions:n.transitions},t.updateQueue.baseState=r,t.memoizedState=r,t.flags&256){s=Ds(Error(R(423)),t),t=om(e,t,i,a,s);break e}else if(i!==s){s=Ds(Error(R(424)),t),t=om(e,t,i,a,s);break e}else for(qt=ai(t.stateNode.containerInfo.firstChild),Dt=t,qe=!0,Zt=null,a=Ty(t,null,i,a),t.child=a;a;)a.flags=a.flags&-3|4096,a=a.sibling;else{if(Ts(),i===s){t=Ra(e,t,a);break e}rt(e,t,i,a)}t=t.child}return t;case 5:return Dy(t),e===null&&wu(t),i=t.type,s=t.pendingProps,r=e!==null?e.memoizedProps:null,n=s.children,gu(i,s)?n=null:r!==null&&gu(i,r)&&(t.flags|=32),Jy(e,t),rt(e,t,n,a),t.child;case 6:return e===null&&wu(t),null;case 13:return Yy(e,t,a);case 4:return Kc(t,t.stateNode.containerInfo),i=t.pendingProps,e===null?t.child=ks(t,null,i,a):rt(e,t,i,a),t.child;case 11:return i=t.type,s=t.pendingProps,s=t.elementType===i?s:Qt(i,s),sm(e,t,i,s,a);case 7:return rt(e,t,t.pendingProps,a),t.child;case 8:return rt(e,t,t.pendingProps.children,a),t.child;case 12:return rt(e,t,t.pendingProps.children,a),t.child;case 10:e:{if(i=t.type._context,s=t.pendingProps,r=t.memoizedProps,n=s.value,ge(vo,i._currentValue),i._currentValue=n,r!==null)if(ia(r.value,n)){if(r.children===s.children&&!ht.current){t=Ra(e,t,a);break e}}else for(r=t.child,r!==null&&(r.return=t);r!==null;){var o=r.dependencies;if(o!==null){n=r.child;for(var l=o.firstContext;l!==null;){if(l.context===i){if(r.tag===1){l=Ca(-1,a&-a),l.tag=2;var d=r.updateQueue;if(d!==null){d=d.shared;var c=d.pending;c===null?l.next=l:(l.next=c.next,c.next=l),d.pending=l}}r.lanes|=a,l=r.alternate,l!==null&&(l.lanes|=a),Tu(r.return,a,t),o.lanes|=a;break}l=l.next}}else if(r.tag===10)n=r.type===t.type?null:r.child;else if(r.tag===18){if(n=r.return,n===null)throw Error(R(341));n.lanes|=a,o=n.alternate,o!==null&&(o.lanes|=a),Tu(n,a,t),n=r.sibling}else n=r.child;if(n!==null)n.return=r;else for(n=r;n!==null;){if(n===t){n=null;break}if(r=n.sibling,r!==null){r.return=n.return,n=r;break}n=n.return}r=n}rt(e,t,s.children,a),t=t.child}return t;case 9:return s=t.type,i=t.pendingProps.children,gs(t,a),s=Ut(s),i=i(s),t.flags|=1,rt(e,t,i,a),t.child;case 14:return i=t.type,s=Qt(i,t.pendingProps),s=Qt(i.type,s),rm(e,t,i,s,a);case 15:return Gy(e,t,t.type,t.pendingProps,a);case 17:return i=t.type,s=t.pendingProps,s=t.elementType===i?s:Qt(i,s),Hn(e,t),t.tag=1,gt(i)?(e=!0,go(t)):e=!1,gs(t,a),Ky(t,i,s),qu(t,i,s,a),Cu(null,t,i,!0,e,a);case 19:return Zy(e,t,a);case 22:return Xy(e,t,a)}throw Error(R(156,t.tag))};function h_(e,t){return zg(e,t)}function v1(e,t,a,i){this.tag=e,this.key=a,this.sibling=this.child=this.return=this.stateNode=this.type=this.elementType=null,this.index=0,this.ref=null,this.pendingProps=t,this.dependencies=this.memoizedState=this.updateQueue=this.memoizedProps=null,this.mode=i,this.subtreeFlags=this.flags=0,this.deletions=null,this.childLanes=this.lanes=0,this.alternate=null}function zt(e,t,a,i){return new v1(e,t,a,i)}function np(e){return e=e.prototype,!(!e||!e.isReactComponent)}function b1(e){if(typeof e=="function")return np(e)?1:0;if(e!=null){if(e=e.$$typeof,e===Dc)return 11;if(e===xc)return 14}return 2}function ni(e,t){var a=e.alternate;return a===null?(a=zt(e.tag,t,e.key,e.mode),a.elementType=e.elementType,a.type=e.type,a.stateNode=e.stateNode,a.alternate=e,e.alternate=a):(a.pendingProps=t,a.type=e.type,a.flags=0,a.subtreeFlags=0,a.deletions=null),a.flags=e.flags&14680064,a.childLanes=e.childLanes,a.lanes=e.lanes,a.child=e.child,a.memoizedProps=e.memoizedProps,a.memoizedState=e.memoizedState,a.updateQueue=e.updateQueue,t=e.dependencies,a.dependencies=t===null?null:{lanes:t.lanes,firstContext:t.firstContext},a.sibling=e.sibling,a.index=e.index,a.ref=e.ref,a}function Jn(e,t,a,i,s,r){var n=2;if(i=e,typeof e=="function")np(e)&&(n=1);else if(typeof e=="string")n=5;else e:switch(e){case es:return Ri(a.children,s,r,t);case qc:n=8,s|=8;break;case Gd:return e=zt(12,a,t,s|2),e.elementType=Gd,e.lanes=r,e;case Xd:return e=zt(13,a,t,s),e.elementType=Xd,e.lanes=r,e;case Jd:return e=zt(19,a,t,s),e.elementType=Jd,e.lanes=r,e;case Dg:return Yo(a,s,r,t);default:if(typeof e=="object"&&e!==null)switch(e.$$typeof){case kg:n=10;break e;case qg:n=9;break e;case Dc:n=11;break e;case xc:n=14;break e;case Ka:n=16,i=null;break e}throw Error(R(130,e==null?e:typeof e,""))}return t=zt(n,a,t,s),t.elementType=e,t.type=i,t.lanes=r,t}function Ri(e,t,a,i){return e=zt(7,e,i,t),e.lanes=a,e}function Yo(e,t,a,i){return e=zt(22,e,i,t),e.elementType=Dg,e.lanes=a,e.stateNode={isHidden:!1},e}function qd(e,t,a){return e=zt(6,e,null,t),e.lanes=a,e}function Dd(e,t,a){return t=zt(4,e.children!==null?e.children:[],e.key,t),t.lanes=a,t.stateNode={containerInfo:e.containerInfo,pendingChildren:null,implementation:e.implementation},t}function w1(e,t,a,i,s){this.tag=t,this.containerInfo=e,this.finishedWork=this.pingCache=this.current=this.pendingChildren=null,this.timeoutHandle=-1,this.callbackNode=this.pendingContext=this.context=null,this.callbackPriority=0,this.eventTimes=rd(0),this.expirationTimes=rd(-1),this.entangledLanes=this.finishedLanes=this.mutableReadLanes=this.expiredLanes=this.pingedLanes=this.suspendedLanes=this.pendingLanes=0,this.entanglements=rd(0),this.identifierPrefix=i,this.onRecoverableError=s,this.mutableSourceEagerHydrationData=null}function op(e,t,a,i,s,r,n,o,l){return e=new w1(e,t,a,o,l),t===1?(t=1,r===!0&&(t|=8)):t=0,r=zt(3,null,null,t),e.current=r,r.stateNode=e,r.memoizedState={element:i,isDehydrated:a,cache:null,transitions:null,pendingSuspenseBoundaries:null},Uc(r),e}function T1(e,t,a){var i=3"u"||typeof __REACT_DEVTOOLS_GLOBAL_HOOK__.checkDCE!="function"))try{__REACT_DEVTOOLS_GLOBAL_HOOK__.checkDCE(v_)}catch(e){console.error(e)}}v_(),vg.exports=Pt;var cp=vg.exports;const Sn=gc(cp);var vm=cp;Vd.createRoot=vm.createRoot,Vd.hydrateRoot=vm.hydrateRoot;/** + * @remix-run/router v1.21.0 + * + * Copyright (c) Remix Software Inc. + * + * This source code is licensed under the MIT license found in the + * LICENSE.md file in the root directory of this source tree. + * + * @license MIT + */function Lr(){return Lr=Object.assign?Object.assign.bind():function(e){for(var t=1;t"u")throw new Error(t)}function b_(e,t){if(!e){typeof console<"u"&&console.warn(t);try{throw new Error(t)}catch{}}}function S1(){return Math.random().toString(36).substr(2,8)}function wm(e,t){return{usr:e.state,key:e.key,idx:t}}function Mu(e,t,a,i){return a===void 0&&(a=null),Lr({pathname:typeof e=="string"?e:e.pathname,search:"",hash:""},typeof t=="string"?Es(t):t,{state:a,key:t&&t.key||i||S1()})}function Ao(e){let{pathname:t="/",search:a="",hash:i=""}=e;return a&&a!=="?"&&(t+=a.charAt(0)==="?"?a:"?"+a),i&&i!=="#"&&(t+=i.charAt(0)==="#"?i:"#"+i),t}function Es(e){let t={};if(e){let a=e.indexOf("#");a>=0&&(t.hash=e.substr(a),e=e.substr(0,a));let i=e.indexOf("?");i>=0&&(t.search=e.substr(i),e=e.substr(0,i)),e&&(t.pathname=e)}return t}function P1(e,t,a,i){i===void 0&&(i={});let{window:s=document.defaultView,v5Compat:r=!1}=i,n=s.history,o=Ya.Pop,l=null,d=c();d==null&&(d=0,n.replaceState(Lr({},n.state,{idx:d}),""));function c(){return(n.state||{idx:null}).idx}function p(){o=Ya.Pop;let T=c(),y=T==null?null:T-d;d=T,l&&l({action:o,location:h.location,delta:y})}function u(T,y){o=Ya.Push;let g=Mu(h.location,T,y);a&&a(g,T),d=c()+1;let _=wm(g,d),w=h.createHref(g);try{n.pushState(_,"",w)}catch(D){if(D instanceof DOMException&&D.name==="DataCloneError")throw D;s.location.assign(w)}r&&l&&l({action:o,location:h.location,delta:1})}function v(T,y){o=Ya.Replace;let g=Mu(h.location,T,y);a&&a(g,T),d=c();let _=wm(g,d),w=h.createHref(g);n.replaceState(_,"",w),r&&l&&l({action:o,location:h.location,delta:0})}function f(T){let y=s.location.origin!=="null"?s.location.origin:s.location.href,g=typeof T=="string"?T:Ao(T);return g=g.replace(/ $/,"%20"),Le(y,"No window.location.(origin|href) available to create URL for href: "+g),new URL(g,y)}let h={get action(){return o},get location(){return e(s,n)},listen(T){if(l)throw new Error("A history only accepts one active listener");return s.addEventListener(bm,p),l=T,()=>{s.removeEventListener(bm,p),l=null}},createHref(T){return t(s,T)},createURL:f,encodeLocation(T){let y=f(T);return{pathname:y.pathname,search:y.search,hash:y.hash}},push:u,replace:v,go(T){return n.go(T)}};return h}var Tm;(function(e){e.data="data",e.deferred="deferred",e.redirect="redirect",e.error="error"})(Tm||(Tm={}));function A1(e,t,a){return a===void 0&&(a="/"),R1(e,t,a,!1)}function R1(e,t,a,i){let s=typeof t=="string"?Es(t):t,r=pp(s.pathname||"/",a);if(r==null)return null;let n=w_(e);I1(n);let o=null;for(let l=0;o==null&&l{let l={relativePath:o===void 0?r.path||"":o,caseSensitive:r.caseSensitive===!0,childrenIndex:n,route:r};l.relativePath.startsWith("/")&&(Le(l.relativePath.startsWith(i),'Absolute route path "'+l.relativePath+'" nested under path '+('"'+i+'" is not valid. An absolute child route path ')+"must start with the combined path of all its parent routes."),l.relativePath=l.relativePath.slice(i.length));let d=oi([i,l.relativePath]),c=a.concat(l);r.children&&r.children.length>0&&(Le(r.index!==!0,"Index routes must not have child routes. Please remove "+('all child routes from route path "'+d+'".')),w_(r.children,t,c,d)),!(r.path==null&&!r.index)&&t.push({path:d,score:M1(d,r.index),routesMeta:c})};return e.forEach((r,n)=>{var o;if(r.path===""||!((o=r.path)!=null&&o.includes("?")))s(r,n);else for(let l of T_(r.path))s(r,n,l)}),t}function T_(e){let t=e.split("/");if(t.length===0)return[];let[a,...i]=t,s=a.endsWith("?"),r=a.replace(/\?$/,"");if(i.length===0)return s?[r,""]:[r];let n=T_(i.join("/")),o=[];return o.push(...n.map(l=>l===""?r:[r,l].join("/"))),s&&o.push(...n),o.map(l=>e.startsWith("/")&&l===""?"/":l)}function I1(e){e.sort((t,a)=>t.score!==a.score?a.score-t.score:L1(t.routesMeta.map(i=>i.childrenIndex),a.routesMeta.map(i=>i.childrenIndex)))}const N1=/^:[\w-]+$/,j1=3,E1=2,F1=1,B1=10,O1=-2,km=e=>e==="*";function M1(e,t){let a=e.split("/"),i=a.length;return a.some(km)&&(i+=O1),t&&(i+=E1),a.filter(s=>!km(s)).reduce((s,r)=>s+(N1.test(r)?j1:r===""?F1:B1),i)}function L1(e,t){return e.length===t.length&&e.slice(0,-1).every((i,s)=>i===t[s])?e[e.length-1]-t[t.length-1]:0}function W1(e,t,a){a===void 0&&(a=!1);let{routesMeta:i}=e,s={},r="/",n=[];for(let o=0;o{let{paramName:u,isOptional:v}=c;if(u==="*"){let h=o[p]||"";n=r.slice(0,r.length-h.length).replace(/(.)\/+$/,"$1")}const f=o[p];return v&&!f?d[u]=void 0:d[u]=(f||"").replace(/%2F/g,"/"),d},{}),pathname:r,pathnameBase:n,pattern:e}}function z1(e,t,a){t===void 0&&(t=!1),a===void 0&&(a=!0),b_(e==="*"||!e.endsWith("*")||e.endsWith("/*"),'Route path "'+e+'" will be treated as if it were '+('"'+e.replace(/\*$/,"/*")+'" because the `*` character must ')+"always follow a `/` in the pattern. To get rid of this warning, "+('please change the route path to "'+e.replace(/\*$/,"/*")+'".'));let i=[],s="^"+e.replace(/\/*\*?$/,"").replace(/^\/*/,"/").replace(/[\\.*+^${}|()[\]]/g,"\\$&").replace(/\/:([\w-]+)(\?)?/g,(n,o,l)=>(i.push({paramName:o,isOptional:l!=null}),l?"/?([^\\/]+)?":"/([^\\/]+)"));return e.endsWith("*")?(i.push({paramName:"*"}),s+=e==="*"||e==="/*"?"(.*)$":"(?:\\/(.+)|\\/*)$"):a?s+="\\/*$":e!==""&&e!=="/"&&(s+="(?:(?=\\/|$))"),[new RegExp(s,t?void 0:"i"),i]}function $1(e){try{return e.split("/").map(t=>decodeURIComponent(t).replace(/\//g,"%2F")).join("/")}catch(t){return b_(!1,'The URL path "'+e+'" could not be decoded because it is is a malformed URL segment. This is probably due to a bad percent '+("encoding ("+t+").")),e}}function pp(e,t){if(t==="/")return e;if(!e.toLowerCase().startsWith(t.toLowerCase()))return null;let a=t.endsWith("/")?t.length-1:t.length,i=e.charAt(a);return i&&i!=="/"?null:e.slice(a)||"/"}function U1(e,t){t===void 0&&(t="/");let{pathname:a,search:i="",hash:s=""}=typeof e=="string"?Es(e):e;return{pathname:a?a.startsWith("/")?a:K1(a,t):t,search:G1(i),hash:X1(s)}}function K1(e,t){let a=t.replace(/\/+$/,"").split("/");return e.split("/").forEach(s=>{s===".."?a.length>1&&a.pop():s!=="."&&a.push(s)}),a.length>1?a.join("/"):"/"}function xd(e,t,a,i){return"Cannot include a '"+e+"' character in a manually specified "+("`to."+t+"` field ["+JSON.stringify(i)+"]. Please separate it out to the ")+("`to."+a+"` field. Alternatively you may provide the full path as ")+'a string in and the router will parse it for you.'}function V1(e){return e.filter((t,a)=>a===0||t.route.path&&t.route.path.length>0)}function k_(e,t){let a=V1(e);return t?a.map((i,s)=>s===a.length-1?i.pathname:i.pathnameBase):a.map(i=>i.pathnameBase)}function q_(e,t,a,i){i===void 0&&(i=!1);let s;typeof e=="string"?s=Es(e):(s=Lr({},e),Le(!s.pathname||!s.pathname.includes("?"),xd("?","pathname","search",s)),Le(!s.pathname||!s.pathname.includes("#"),xd("#","pathname","hash",s)),Le(!s.search||!s.search.includes("#"),xd("#","search","hash",s)));let r=e===""||s.pathname==="",n=r?"/":s.pathname,o;if(n==null)o=a;else{let p=t.length-1;if(!i&&n.startsWith("..")){let u=n.split("/");for(;u[0]==="..";)u.shift(),p-=1;s.pathname=u.join("/")}o=p>=0?t[p]:"/"}let l=U1(s,o),d=n&&n!=="/"&&n.endsWith("/"),c=(r||n===".")&&a.endsWith("/");return!l.pathname.endsWith("/")&&(d||c)&&(l.pathname+="/"),l}const oi=e=>e.join("/").replace(/\/\/+/g,"/"),H1=e=>e.replace(/\/+$/,"").replace(/^\/*/,"/"),G1=e=>!e||e==="?"?"":e.startsWith("?")?e:"?"+e,X1=e=>!e||e==="#"?"":e.startsWith("#")?e:"#"+e;function J1(e){return e!=null&&typeof e.status=="number"&&typeof e.statusText=="string"&&typeof e.internal=="boolean"&&"data"in e}const D_=["post","put","patch","delete"];new Set(D_);const Q1=["get",...D_];new Set(Q1);/** + * React Router v6.28.0 + * + * Copyright (c) Remix Software Inc. + * + * This source code is licensed under the MIT license found in the + * LICENSE.md file in the root directory of this source tree. + * + * @license MIT + */function Wr(){return Wr=Object.assign?Object.assign.bind():function(e){for(var t=1;t{o.current=!0}),b.useCallback(function(d,c){if(c===void 0&&(c={}),!o.current)return;if(typeof d=="number"){i.go(d);return}let p=q_(d,JSON.parse(n),r,c.relative==="path");e==null&&t!=="/"&&(p.pathname=p.pathname==="/"?t:oi([t,p.pathname])),(c.replace?i.replace:i.push)(p,c.state,c)},[t,i,n,r,e])}function aT(){let{matches:e}=b.useContext(fi),t=e[e.length-1];return t?t.params:{}}function S_(e,t){let{relative:a}=t===void 0?{}:t,{future:i}=b.useContext(Wi),{matches:s}=b.useContext(fi),{pathname:r}=sl(),n=JSON.stringify(k_(s,i.v7_relativeSplatPath));return b.useMemo(()=>q_(e,JSON.parse(n),r,a==="path"),[e,n,r,a])}function iT(e,t){return sT(e,t)}function sT(e,t,a,i){Yr()||Le(!1);let{navigator:s}=b.useContext(Wi),{matches:r}=b.useContext(fi),n=r[r.length-1],o=n?n.params:{};n&&n.pathname;let l=n?n.pathnameBase:"/";n&&n.route;let d=sl(),c;if(t){var p;let T=typeof t=="string"?Es(t):t;l==="/"||(p=T.pathname)!=null&&p.startsWith(l)||Le(!1),c=T}else c=d;let u=c.pathname||"/",v=u;if(l!=="/"){let T=l.replace(/^\//,"").split("/");v="/"+u.replace(/^\//,"").split("/").slice(T.length).join("/")}let f=A1(e,{pathname:v}),h=dT(f&&f.map(T=>Object.assign({},T,{params:Object.assign({},o,T.params),pathname:oi([l,s.encodeLocation?s.encodeLocation(T.pathname).pathname:T.pathname]),pathnameBase:T.pathnameBase==="/"?l:oi([l,s.encodeLocation?s.encodeLocation(T.pathnameBase).pathname:T.pathnameBase])})),r,a,i);return t&&h?b.createElement(il.Provider,{value:{location:Wr({pathname:"/",search:"",hash:"",state:null,key:"default"},c),navigationType:Ya.Pop}},h):h}function rT(){let e=fT(),t=J1(e)?e.status+" "+e.statusText:e instanceof Error?e.message:JSON.stringify(e),a=e instanceof Error?e.stack:null,s={padding:"0.5rem",backgroundColor:"rgba(200,200,200, 0.5)"},r=null;return b.createElement(b.Fragment,null,b.createElement("h2",null,"Unexpected Application Error!"),b.createElement("h3",{style:{fontStyle:"italic"}},t),a?b.createElement("pre",{style:s},a):null,r)}const nT=b.createElement(rT,null);class oT extends b.Component{constructor(t){super(t),this.state={location:t.location,revalidation:t.revalidation,error:t.error}}static getDerivedStateFromError(t){return{error:t}}static getDerivedStateFromProps(t,a){return a.location!==t.location||a.revalidation!=="idle"&&t.revalidation==="idle"?{error:t.error,location:t.location,revalidation:t.revalidation}:{error:t.error!==void 0?t.error:a.error,location:a.location,revalidation:t.revalidation||a.revalidation}}componentDidCatch(t,a){console.error("React Router caught the following error during render",t,a)}render(){return this.state.error!==void 0?b.createElement(fi.Provider,{value:this.props.routeContext},b.createElement(x_.Provider,{value:this.state.error,children:this.props.component})):this.props.children}}function lT(e){let{routeContext:t,match:a,children:i}=e,s=b.useContext(fp);return s&&s.static&&s.staticContext&&(a.route.errorElement||a.route.ErrorBoundary)&&(s.staticContext._deepestRenderedBoundaryId=a.route.id),b.createElement(fi.Provider,{value:t},i)}function dT(e,t,a,i){var s;if(t===void 0&&(t=[]),a===void 0&&(a=null),i===void 0&&(i=null),e==null){var r;if(!a)return null;if(a.errors)e=a.matches;else if((r=i)!=null&&r.v7_partialHydration&&t.length===0&&!a.initialized&&a.matches.length>0)e=a.matches;else return null}let n=e,o=(s=a)==null?void 0:s.errors;if(o!=null){let c=n.findIndex(p=>p.route.id&&(o==null?void 0:o[p.route.id])!==void 0);c>=0||Le(!1),n=n.slice(0,Math.min(n.length,c+1))}let l=!1,d=-1;if(a&&i&&i.v7_partialHydration)for(let c=0;c=0?n=n.slice(0,d+1):n=[n[0]];break}}}return n.reduceRight((c,p,u)=>{let v,f=!1,h=null,T=null;a&&(v=o&&p.route.id?o[p.route.id]:void 0,h=p.route.errorElement||nT,l&&(d<0&&u===0?(hT("route-fallback",!1),f=!0,T=null):d===u&&(f=!0,T=p.route.hydrateFallbackElement||null)));let y=t.concat(n.slice(0,u+1)),g=()=>{let _;return v?_=h:f?_=T:p.route.Component?_=b.createElement(p.route.Component,null):p.route.element?_=p.route.element:_=c,b.createElement(lT,{match:p,routeContext:{outlet:c,matches:y,isDataRoute:a!=null},children:_})};return a&&(p.route.ErrorBoundary||p.route.errorElement||u===0)?b.createElement(oT,{location:a.location,revalidation:a.revalidation,component:h,error:v,children:g(),routeContext:{outlet:null,matches:y,isDataRoute:!0}}):g()},null)}var P_=function(e){return e.UseBlocker="useBlocker",e.UseRevalidator="useRevalidator",e.UseNavigateStable="useNavigate",e}(P_||{}),Ro=function(e){return e.UseBlocker="useBlocker",e.UseLoaderData="useLoaderData",e.UseActionData="useActionData",e.UseRouteError="useRouteError",e.UseNavigation="useNavigation",e.UseRouteLoaderData="useRouteLoaderData",e.UseMatches="useMatches",e.UseRevalidator="useRevalidator",e.UseNavigateStable="useNavigate",e.UseRouteId="useRouteId",e}(Ro||{});function uT(e){let t=b.useContext(fp);return t||Le(!1),t}function cT(e){let t=b.useContext(Y1);return t||Le(!1),t}function pT(e){let t=b.useContext(fi);return t||Le(!1),t}function A_(e){let t=pT(),a=t.matches[t.matches.length-1];return a.route.id||Le(!1),a.route.id}function fT(){var e;let t=b.useContext(x_),a=cT(Ro.UseRouteError),i=A_(Ro.UseRouteError);return t!==void 0?t:(e=a.errors)==null?void 0:e[i]}function mT(){let{router:e}=uT(P_.UseNavigateStable),t=A_(Ro.UseNavigateStable),a=b.useRef(!1);return C_(()=>{a.current=!0}),b.useCallback(function(s,r){r===void 0&&(r={}),a.current&&(typeof s=="number"?e.navigate(s):e.navigate(s,Wr({fromRouteId:t},r)))},[e,t])}const Dm={};function hT(e,t,a){!t&&!Dm[e]&&(Dm[e]=!0)}const xm={};function gT(e,t){xm[t]||(xm[t]=!0,console.warn(t))}const Ki=(e,t,a)=>gT(e,"⚠️ React Router Future Flag Warning: "+t+". "+("You can use the `"+e+"` future flag to opt-in early. ")+("For more information, see "+a+"."));function yT(e,t){e!=null&&e.v7_startTransition||Ki("v7_startTransition","React Router will begin wrapping state updates in `React.startTransition` in v7","https://reactrouter.com/v6/upgrading/future#v7_starttransition"),!(e!=null&&e.v7_relativeSplatPath)&&(!t||!t.v7_relativeSplatPath)&&Ki("v7_relativeSplatPath","Relative route resolution within Splat routes is changing in v7","https://reactrouter.com/v6/upgrading/future#v7_relativesplatpath"),t&&(t.v7_fetcherPersist||Ki("v7_fetcherPersist","The persistence behavior of fetchers is changing in v7","https://reactrouter.com/v6/upgrading/future#v7_fetcherpersist"),t.v7_normalizeFormMethod||Ki("v7_normalizeFormMethod","Casing of `formMethod` fields is being normalized to uppercase in v7","https://reactrouter.com/v6/upgrading/future#v7_normalizeformmethod"),t.v7_partialHydration||Ki("v7_partialHydration","`RouterProvider` hydration behavior is changing in v7","https://reactrouter.com/v6/upgrading/future#v7_partialhydration"),t.v7_skipActionErrorRevalidation||Ki("v7_skipActionErrorRevalidation","The revalidation behavior after 4xx/5xx `action` responses is changing in v7","https://reactrouter.com/v6/upgrading/future#v7_skipactionerrorrevalidation"))}function Lu(e){Le(!1)}function _T(e){let{basename:t="/",children:a=null,location:i,navigationType:s=Ya.Pop,navigator:r,static:n=!1,future:o}=e;Yr()&&Le(!1);let l=t.replace(/^\/*/,"/"),d=b.useMemo(()=>({basename:l,navigator:r,static:n,future:Wr({v7_relativeSplatPath:!1},o)}),[l,o,r,n]);typeof i=="string"&&(i=Es(i));let{pathname:c="/",search:p="",hash:u="",state:v=null,key:f="default"}=i,h=b.useMemo(()=>{let T=pp(c,l);return T==null?null:{location:{pathname:T,search:p,hash:u,state:v,key:f},navigationType:s}},[l,c,p,u,v,f,s]);return h==null?null:b.createElement(Wi.Provider,{value:d},b.createElement(il.Provider,{children:a,value:h}))}function vT(e){let{children:t,location:a}=e;return iT(Wu(t),a)}new Promise(()=>{});function Wu(e,t){t===void 0&&(t=[]);let a=[];return b.Children.forEach(e,(i,s)=>{if(!b.isValidElement(i))return;let r=[...t,s];if(i.type===b.Fragment){a.push.apply(a,Wu(i.props.children,r));return}i.type!==Lu&&Le(!1),!i.props.index||!i.props.children||Le(!1);let n={id:i.props.id||r.join("-"),caseSensitive:i.props.caseSensitive,element:i.props.element,Component:i.props.Component,index:i.props.index,path:i.props.path,loader:i.props.loader,action:i.props.action,errorElement:i.props.errorElement,ErrorBoundary:i.props.ErrorBoundary,hasErrorBoundary:i.props.ErrorBoundary!=null||i.props.errorElement!=null,shouldRevalidate:i.props.shouldRevalidate,handle:i.props.handle,lazy:i.props.lazy};i.props.children&&(n.children=Wu(i.props.children,r)),a.push(n)}),a}/** + * React Router DOM v6.28.0 + * + * Copyright (c) Remix Software Inc. + * + * This source code is licensed under the MIT license found in the + * LICENSE.md file in the root directory of this source tree. + * + * @license MIT + */function zu(){return zu=Object.assign?Object.assign.bind():function(e){for(var t=1;t=0)&&(a[s]=e[s]);return a}function wT(e){return!!(e.metaKey||e.altKey||e.ctrlKey||e.shiftKey)}function TT(e,t){return e.button===0&&(!t||t==="_self")&&!wT(e)}const kT=["onClick","relative","reloadDocument","replace","state","target","to","preventScrollReset","viewTransition"],qT="6";try{window.__reactRouterVersion=qT}catch{}const DT="startTransition",Cm=so[DT];function xT(e){let{basename:t,children:a,future:i,window:s}=e,r=b.useRef();r.current==null&&(r.current=C1({window:s,v5Compat:!0}));let n=r.current,[o,l]=b.useState({action:n.action,location:n.location}),{v7_startTransition:d}=i||{},c=b.useCallback(p=>{d&&Cm?Cm(()=>l(p)):l(p)},[l,d]);return b.useLayoutEffect(()=>n.listen(c),[n,c]),b.useEffect(()=>yT(i),[i]),b.createElement(_T,{basename:t,children:a,location:o.location,navigationType:o.action,navigator:n,future:i})}const CT=typeof window<"u"&&typeof window.document<"u"&&typeof window.document.createElement<"u",ST=/^(?:[a-z][a-z0-9+.-]*:|\/\/)/i,Bi=b.forwardRef(function(t,a){let{onClick:i,relative:s,reloadDocument:r,replace:n,state:o,target:l,to:d,preventScrollReset:c,viewTransition:p}=t,u=bT(t,kT),{basename:v}=b.useContext(Wi),f,h=!1;if(typeof d=="string"&&ST.test(d)&&(f=d,CT))try{let _=new URL(window.location.href),w=d.startsWith("//")?new URL(_.protocol+d):new URL(d),D=pp(w.pathname,v);w.origin===_.origin&&D!=null?d=D+w.search+w.hash:h=!0}catch{}let T=Z1(d,{relative:s}),y=PT(d,{replace:n,state:o,target:l,preventScrollReset:c,relative:s,viewTransition:p});function g(_){i&&i(_),_.defaultPrevented||y(_)}return b.createElement("a",zu({},u,{href:f||T,onClick:h||r?i:g,ref:a,target:l}))});var Sm;(function(e){e.UseScrollRestoration="useScrollRestoration",e.UseSubmit="useSubmit",e.UseSubmitFetcher="useSubmitFetcher",e.UseFetcher="useFetcher",e.useViewTransitionState="useViewTransitionState"})(Sm||(Sm={}));var Pm;(function(e){e.UseFetcher="useFetcher",e.UseFetchers="useFetchers",e.UseScrollRestoration="useScrollRestoration"})(Pm||(Pm={}));function PT(e,t){let{target:a,replace:i,state:s,preventScrollReset:r,relative:n,viewTransition:o}=t===void 0?{}:t,l=eT(),d=sl(),c=S_(e,{relative:n});return b.useCallback(p=>{if(TT(p,a)){p.preventDefault();let u=i!==void 0?i:Ao(d)===Ao(c);l(e,{replace:u,state:s,preventScrollReset:r,relative:n,viewTransition:o})}},[d,l,c,i,s,a,e,r,n,o])}const AT="/aiida-registry/assets/logo-white-text-16948862.svg",RT="/aiida-registry/assets/MARVEL-32e738c9.png",IT="/aiida-registry/assets/MaX-099f261c.png";const NT={"aiida-QECpWorkChain":{code_home:"https://github.com/rikigigi/aiida-QECpWorkChain",development_status:"beta",entry_point_prefix:"qecpworkchain",pip_url:"git+https://github.com/rikigigi/aiida-QECpWorkChain",name:"aiida-QECpWorkChain",package_name:"aiida_QECpWorkChain",hosted_on:"github.com",metadata:{author:"Riccardo Bertossa",author_email:"rbertoss@sissa.it",version:"0.2.0a0",description:"Car-Parrinello Work Chain with Quantum Espresso. This workchain does a full CP simulation, from the choice of the electronic mass and the timestep, to the choice of the best parallelization options, and then it does the NPT equilibration and a final NVE simulation at the prescribed P and T. Automates as much as possible.",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: GNU General Public License v3 (GPLv3)","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.workflows":{"qecpworkchain.cp":{description:["No description available"],spec:{inputs:[{name:"cp_code",required:!0,valid_types:"Code",info:""},{name:"cp_resources_cg_list",required:!0,valid_types:"List",info:"Same as cp_resources_cp_list but when doing a CG. The CG uses a different amount of resource and can use no band or task group parallelization."},{name:"cp_resources_cp_list",required:!0,valid_types:"List",info:`List of dictionary like the following: +{ + 'resources' : { + 'num_machines' : 2, + 'num_mpiprocs_per_machine' : 48, + }, + 'wallclock' : 3600, + 'queue' : 'queue_name', + 'account': 'account_name', +} +c,porturrently only the first element of the list is used. +'wallclock' is the maximum time that can be requested to the scheduler. This code can decide to ask for less. +`},{name:"ecutwfc",required:!0,valid_types:"Float",info:"wavefunction cutoff (Ry), like in the QE input"},{name:"pseudo_family",required:!0,valid_types:"Str",info:"pseudopotential family to use, as in usual aiida operations"},{name:"pw_code",required:!0,valid_types:"Code",info:"input pw code (used to calculate force ratio)"},{name:"pw_resources_list",required:!0,valid_types:"List",info:"Same as cp_resources_cp_list but for pw.x code."},{name:"structure",required:!0,valid_types:"StructureData, TrajectoryData",info:"Input structure. If a trajectory is given, the workchain will use its last step to start the CG. If velocities are present, they will be used to initialize the simulation. Note that if you use a trajectory, usually kind information (like mass) are not included, so default values will be used. If you want to include kind information or override those provided with the input structure, use the input structure_kinds"},{name:"thermobarostat_points",required:!0,valid_types:"List",info:'List of dicts, each with the format [ { "temperature_K": 1000, "pressure_KBar": 10 , "equilibration_time_ps": 5.0, "thermostat_time_ps": 5.0} ]. The simulation will loop over this list of dictionaries, in the same order, equilibrating for the specified time at the given P,T point. Every point is repeated if the average T and P are not within the specified ranges'},{name:"additional_parameters_cp",required:!1,valid_types:"Dict",info:"parameters that will be included in the settings input of the QE CP plugin. These settings will be added on top of the default one. Same format as plugin input"},{name:"adjust_ionic_mass",required:!1,valid_types:"Bool",info:"Multiply the mass of the ions by the corresponding force ration between the cp forces and pw forces -- that is less than 1. Note that averages of static properties do not depend on the ionic masses."},{name:"benchmark_emass_dt_walltime_s",required:!1,valid_types:"Float",info:"same as benchmark_parallel_walltime_s but for dermining the best electronic mass and timestep."},{name:"benchmark_parallel_walltime_s",required:!1,valid_types:"Float",info:"time requested to the scheduler during the test for finding the best parallelization parameters."},{name:"cmdline_cp",required:!1,valid_types:"List, NoneType",info:"additional command line parameters of the cp verlet caclulations only (for example parallelization options)"},{name:"default_nose_frequency",required:!1,valid_types:"Float",info:"default nose frequency when a frequency cannot be estimated from the vibrational spectrum"},{name:"dt",required:!1,valid_types:"Float, NoneType",info:"timestep in atomic units, if not automatically chosen."},{name:"dt_start_stop_step",required:!1,valid_types:"List",info:"list of timesteps to try. Timesteps are changed to better integrate the equation of motion. When a new electronic mass is selected by this workchain timesteps are automatically adjusted."},{name:"emass",required:!1,valid_types:"Float, NoneType",info:"electronic mass, atomic mass units, if not automatically chosen"},{name:"emass_list",required:!1,valid_types:"List",info:"list of electronic masses to try. The emass is selected in order to satisfy the requested CP/DFT force ratio."},{name:"initial_atomic_velocities_A_ps",required:!1,valid_types:"ArrayData, NoneType",info:"optional input initial velocities in angstrom over picoseconds"},{name:"max_slope_const",required:!1,valid_types:"Float",info:"max slope in K/ps of the constant of motion linear fit."},{name:"max_slope_ekinc",required:!1,valid_types:"Float",info:"max slope in K/ps of the ekinc linear fit. If not satisfied try to change emass"},{name:"max_slope_min_emass",required:!1,valid_types:"Float",info:"minimum possible value of electronic mass that can be set by the max_slope correction routine. Will not go lower than that."},{name:"max_slope_min_ps",required:!1,valid_types:"Float",info:"minimum required lenght in ps of the last trajectory to do the linear fit on ekinc and const of motion"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"min_traj_steps_vdos",required:!1,valid_types:"Int",info:"minimum number of steps to consider the calculated vibrational spectrum maximum valid, to set the thermostat frequency"},{name:"minimum_nose_frequency",required:!1,valid_types:"Float",info:"minimum nose frequency: if the frequency estimated from the vibrational spectrum is lower than this value, this value is used"},{name:"nstep_initial_cg",required:!1,valid_types:"Int",info:"At the beginning of the simulation the CP algorithm is not used. This is the number of steps to do using Born-Oppenheimer molecular dynamics algorithm with a conjugate gradient minimization of the electronic ground state."},{name:"nstep_parallel_test",required:!1,valid_types:"Int",info:"the benchmark simulations will be that long, if performed"},{name:"number_of_pw_per_trajectory",required:!1,valid_types:"Int",info:"Number of pw submitted for every trajectory during calculation of force ratio."},{name:"nve_required_picoseconds",required:!1,valid_types:"Float",info:"The equilibrated NVE simulation will last at least this number of picoseconds. How much picoseconds do you want?"},{name:"pressure_tolerance",required:!1,valid_types:"Float",info:"Pressure tolerance in kBar used to say if the npt is equilibrated. If not setted, use the standard deviation of the P time series"},{name:"skip_emass_dt_test",required:!1,valid_types:"Bool",info:""},{name:"skip_parallel_test",required:!1,valid_types:"Bool",info:"do not run run benchmarks to discover a good internal Quantum Espresso parallelization scheme for the current system"},{name:"skip_thermobarostat",required:!1,valid_types:"Bool",info:""},{name:"structure_kinds",required:!1,valid_types:"List, NoneType",info:'These kinds will be used to override or set the masses of the various atomic types. Note that the workflow, if skip_emass_dt_test is True, will calculate the ratio between cp forces and pw forces and adjust the provided masses automatically according to this ratio. So if you provide this input, make sure to set skip_emass_dt_test to True and set also the inputs emass and dt, or "bad things can happen"'},{name:"target_force_ratio",required:!1,valid_types:"Float",info:"The forces calculated by the Car-Parrinello method are affected by two types of error: one is due to the oscillations of the electrons around the DFT energy minimum, and the second is due to the finite mass of the electronic fluid that produces a _sistematic_ error in the forces, as if the electrons add mass to the ionic core. This second kind of error is can be controlled by this parameter, that tries to adjust the electronic mass to obtain the desidered ratio between CP forces and true DFT forces. Then you may want to modify the ionic mass to correct the leading factor of this error."},{name:"temperature_tolerance",required:!1,valid_types:"Float",info:"Temperature tolerance in K used to say if the npt is equilibrated. If not setted, use the standard deviation of the T time series"},{name:"tempw_initial_random",required:!1,valid_types:"Float, NoneType",info:"If provided, sets the initial temperature when randomly initializing the starting velocities."}],outputs:[{name:"dt",required:!0,valid_types:"Float",info:"The best integration timestep found"},{name:"emass",required:!0,valid_types:"Float",info:"The best electronic mass found"},{name:"full_traj",required:!0,valid_types:"TrajectoryData",info:"The full merged trajectory of all the simulations"},{name:"nve_prod_traj",required:!0,valid_types:"TrajectoryData",info:"The trajectory of the last NVE simulation"},{name:"cmdline_cp",required:!1,valid_types:"List",info:"The best command line parameters for the cp code"},{name:"kinds",required:!1,valid_types:"List",info:"The kinds for the structure with the modified ionic masses"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"The initial cg steps failed. I cannot start to work."},{status:402,message:"Nose-Hoover thermostat failed."},{status:403,message:"Final cg after Nose-Hoover failed."},{status:404,message:"Error in the NVE simulation"},{status:405,message:"The simulations are calculating very expensive random numbers. There is something wrong (cutoff? metal? boo?)"},{status:406,message:"Wrong input parameters"},{status:407,message:"Parallel test was not succesful, maybe there is something more wrong."},{status:408,message:"Multiple errors in the simulation that cannot fix."},{status:409,message:"This is a bug in the workchain."},{status:410,message:"The initial simulation failed. I cannot start to work. Probably the configuration is not suitable for a molecular dynamics run or the trial integration timestep is too big"}]},class:"aiida_QECpWorkChain.workflow:CpWorkChain"}}},commits_count:6,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/rikigigi/aiida-QECpWorkChain",is_installable:"True"},"aiida-abinit":{code_home:"https://github.com/sponce24/aiida-abinit",entry_point_prefix:"abinit",pip_url:"aiida-abinit",plugin_info:"https://raw.github.com/sponce24/aiida-abinit/master/setup.json",name:"aiida-abinit",package_name:"aiida_abinit",hosted_on:"github.com",metadata:{release_date:"2023-10-04",description:"The AiiDA plugin for ABINIT.",author_email:"Samuel Ponce ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"0.5.0"},aiida_version:">=2.0,<3.0",entry_points:{"aiida.calculations":{abinit:{description:["AiiDA calculation plugin wrapping the abinit executable."],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The k-point mesh or path"},{name:"parameters",required:!0,valid_types:"Dict",info:"The ABINIT input parameters."},{name:"pseudos",required:!0,valid_types:"Psp8Data, JthXmlData",info:"The pseudopotentials."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"A remote folder used for restarts."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Various special settings."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"Various output quantities."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_bands",required:!1,valid_types:"BandsData",info:"Final electronic bands if present."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Final structure of the calculation if present."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:"Trajectory of various output quantities over the calculation if present."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"Calculation did not produce all expected output files."},{status:101,message:"Calculation did not produce the expected `[prefix]o_GSR.nc` output file."},{status:102,message:"Calculation did not produce the expected `[prefix]o_HIST.nc` output file."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"The retrieved folder data node could not be accessed."},{status:210,message:"The retrieved folder did not contain the `stdout` output file."},{status:301,message:"The `stdout` output file could not be read."},{status:302,message:"The `stdout` output file could not be parsed."},{status:303,message:"The `abipy` `EventsParser` reports that the runw as not completed."},{status:304,message:"The output file contains one or more error messages."},{status:305,message:"The output file contains one or more warning messages."},{status:312,message:"The output structure could not be parsed."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:500,message:"The SCF minimization cycle did not converge."},{status:501,message:"The ionic minimization cycle did not converge."}]},class:"aiida_abinit.calculations:AbinitCalculation"}},"aiida.parsers":{abinit:"aiida_abinit.parsers:AbinitParser"},"aiida.workflows":{"abinit.base":{description:["Base Abinit Workchain to perform a DFT calculation. Validates parameters and restart."],spec:{inputs:[{name:"abinit",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"An explicit k-points mesh or list. Either this or `kpoints_distance` must be provided."},{name:"kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"The minimum desired distance in 1/Å between k-points in reciprocal space. The explicit k-point mesh will be generated automatically by a calculation function based on the input structure."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"Various output quantities."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_bands",required:!1,valid_types:"BandsData",info:"Final electronic bands if present."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Final structure of the calculation if present."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:"Trajectory of various output quantities over the calculation if present."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"`pseudos` could not be used to get the necessary pseudos."},{status:202,message:"Neither the `kpoints` nor the `kpoints_distance` input was specified."},{status:203,message:"Neither the `options` nor `automatic_parallelization` input was specified."},{status:204,message:"The `metadata.options` did not specify both `resources.num_machines` and `max_wallclock_seconds`."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_abinit.workflows.base:AbinitBaseWorkChain"}}},commits_count:0,development_status:"beta",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-abinit",is_installable:"True"},"aiida-aenet":{code_home:"https://gitlab.com/lattice737/aiida-aenet",development_status:"planning",entry_point_prefix:"aenet",pip_url:"https://gitlab.com/lattice737/aiida-aenet",name:"aiida-aenet",package_name:"aiida_aenet",hosted_on:"gitlab.com",metadata:{author:"Nicholas Martinez",author_email:"nicholasmartinez@my.unt.edu",version:"0.1.0",description:"AiiDA plugin to construct machine-learning potentials using aenet",classifiers:["Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Development Status :: 0 - Alpha"]},aiida_version:"~=1.2",entry_points:{"aiida.data":{"aenet.algorithm":"aiida_aenet.data.algorithm:AenetAlgorithm","aenet.potential":"aiida_aenet.data.potentials:AenetPotential"},"aiida.calculations":{"aenet.cur":"aiida_aenet.calculations.cur:CurCalculation","aenet.generate":"aiida_aenet.calculations.generate:AenetGenerateCalculation","aenet.predict":"aiida_aenet.calculations.predict:AenetPredictCalculation","aenet.simulate":"aiida_aenet.calculations.simulate:AenetLammpsMdCalculation","aenet.train":"aiida_aenet.calculations.train:AenetTrainCalculation","aenet.transform":"aiida_aenet.calculations.transform:TransformCalculation"},"aiida.parsers":{"aenet.generate":"aiida_aenet.parsers.generate:AenetGenerateParser","aenet.predict":"aiida_aenet.parsers.predict:AenetPredictParser","aenet.simulate":"aiida_aenet.parsers.simulate:AenetLammpsMdParser","aenet.train":"aiida_aenet.parsers.train:AenetTrainParser"},"aiida.workflows":{"aenet.build_reference":"aiida_aenet.workflows.build_reference:BuildReferenceWorkChain","aenet.compare_simulations":"aiida_aenet.workflows.compare_simulations:CompareSimulationsWorkChain","aenet.make_potential":"aiida_aenet.workflows.make_potential:MakePotentialWorkChain","aenet.make_structures":"aiida_aenet.workflows.make_structures:MakeStructuresWorkChain"},"aenet.potentials":{"lammps.ann":"aiida_aenet.data.potentials.lammps:ANN"}},commits_count:0,errors:[],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:6},{colorclass:"brown",text:"Parsers",count:4},{colorclass:"red",text:"Data",count:2},{colorclass:"green",text:"Workflows",count:4},{colorclass:"orange",text:"Other (Aenet potentials)",count:1}],pip_install_cmd:"pip install https://gitlab.com/lattice737/aiida-aenet"},"aiida-aimall":{entry_point_prefix:"aimall",code_home:"https://github.com/kmlefran/aiida-aimall",pip_url:"aiida-aimall",documentation_url:"https://aiida-aimall.readthedocs.io/en/latest/",plugin_info:"https://raw.githubusercontent.com/kmlefran/aiida-aimall/main/pyproject.toml",version_file:"https://raw.githubusercontent.com/kmlefran/aiida-aimall/main/aiida_aimall/__init__.py",name:"aiida-aimall",package_name:"aiida_aimall",hosted_on:"github.com",metadata:{release_date:"2024-09-12",description:"A plugin to interface AIMAll with AiiDA",author_email:"Kevin Lefrancois-Gagnon ",classifiers:["Development Status :: 3 - Alpha","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11"],version:"1.0.1"},aiida_version:">=2.0,<3",entry_points:{"aiida.calculations":{"aimall.aimqb":{description:["AiiDA calculation plugin wrapping the aimqb executable.",""," Attributes:"," parameters (AimqbParameters): command line parameters for the AimqbCalculation"," file (aiida.orm.SinglefileData): the wfx, wfn, or fchk file to be run"," code (aiida.orm.Code): code of the AIMQB executable"," attached_atom_int (aiida.orm.Int): optional integer label of the atom that is attached to the rest of the molecule"," group_atoms (aiida.orm.List): optional integer list of ids of atoms comprising the group for AimqbGroupParser",""," Example:"," ::",""," code = orm.load_code('aimall@localhost')",' AimqbParameters = DataFactory("aimall.aimqb")',' aim_params = AimqbParameters(parameter_dict={"naat": 2, "nproc": 2, "atlaprhocps": True})',' file = SinglefileData("/absolute/path/to/file")'," # Alternatively, if you have the file as a string, you can build the file with:"," # file=SinglefileData(io.BytesIO(file_string.encode()))",' AimqbCalculation = CalculationFactory("aimall.aimqb")'," builder = AimqbCalculation.get_builder()"," builder.parameters = aim_params"," builder.file = file"," builder.code = code",' builder.metadata.options.resources = {"num_machines": 1, "num_mpiprocs_per_machine": 2}'," builder.submit()",""," Note:"," By default, the AimqbBaseParser is used, getting atomic, BCP, and (if applicable) LapRhoCps."," You can opt to use the AimqbGroupParser, which also returns the integrated group properties"," of a group, as well as the atomic graph descriptor of the group. In doing so, you can also"," define the atoms included in the group, which, by convention, defaults to all atoms except atom 2."," You can further specify which atom of the group is the one bonded to the substrate, which defaults to"," atom 1. This is done by providing this to the builder:",""," ::","",' builder.metadata.options.parser_name = "aimall.group"'," builder.attached_atom_int = Int(1)"," builder.group_atoms = List([1,3,4,5,6])"],spec:{inputs:[{name:"file",required:!0,valid_types:"SinglefileData",info:"fchk, wfn, or wfx to run AimQB on"},{name:"parameters",required:!0,valid_types:"AimqbParameters",info:"Command line parameters for aimqb"},{name:"attached_atom_int",required:!1,valid_types:"Int",info:"id # of attached atom for graph descriptor. Defaults to atom 1"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"group_atoms",required:!1,valid_types:"List",info:"Integer ids of atoms in groups to include. e.g. [1,3,4]. Defaults to all atoms in molecule except atom 2"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The computed parameters of an AIMAll calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:210,message:"The retrieved folder did not contain the output file."}]},class:"aiida_aimall.calculations:AimqbCalculation"}},"aiida.data":{"aimall.aimqb":"aiida_aimall.data:AimqbParameters"},"aiida.parsers":{"aimall.base":"aiida_aimall.parsers:AimqbBaseParser","aimall.group":"aiida_aimall.parsers:AimqbGroupParser"},"aiida.workflows":{"aimall.aimreor":"aiida_aimall.workchains.param_parts:AIMAllReorWorkChain","aimall.gausstoaim":"aiida_aimall.workchains.qc_programs:GaussianToAIMWorkChain","aimall.qmtoaim":"aiida_aimall.workchains.qc_programs:QMToAIMWorkChain","aimall.smitogauss":"aiida_aimall.workchains.param_parts:SmilesToGaussianWorkChain","aimall.subparam":"aiida_aimall.workchains.subparam:SubstituentParameterWorkChain","aimall.wfxtoaim":"aiida_aimall.workchains.qc_programs:GenerateWFXToAIMWorkChain"}},commits_count:274,development_status:"alpha",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:6}],pip_install_cmd:"pip install aiida-aimall",is_installable:"True"},"aiida-alloy":{code_home:"https://github.com/DanielMarchand/aiida-alloy",development_status:"beta",entry_point_prefix:"alloy",pip_url:"git+https://github.com/DanielMarchand/aiida-alloy",name:"aiida-alloy",package_name:"aiida_alloy",hosted_on:"github.com",metadata:{author:"The AiiDA developers group",author_email:"",version:"0.1.0a0",description:"Aiida Workflows for Elastic Constants using Quantum Espresso",classifiers:["Programming Language :: Python"]},aiida_version:">=1.0.0a0",entry_points:{"aiida.workflows":{elastic:"aiida_alloy.workflows.ElasticWorkChain:ElasticWorkChain"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-alloy
Collecting git+https://github.com/DanielMarchand/aiida-alloy
+  Cloning https://github.com/DanielMarchand/aiida-alloy to /tmp/pip-req-build-8h1hn0so
+  Running command git clone --filter=blob:none --quiet https://github.com/DanielMarchand/aiida-alloy /tmp/pip-req-build-8h1hn0so
+  Resolved https://github.com/DanielMarchand/aiida-alloy to commit dfae54b29291dfe94d4a5ca136738020f447bc1e
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'error'
+  error: subprocess-exited-with-error
+  
+  × python setup.py egg_info did not run successfully.
+  │ exit code: 1
+  ╰─> [6 lines of output]
+      Traceback (most recent call last):
+        File "", line 2, in 
+        File "", line 34, in 
+        File "/tmp/pip-req-build-8h1hn0so/setup.py", line 20, in 
+          long_description=open('README.md').read(),
+      FileNotFoundError: [Errno 2] No such file or directory: 'README.md'
+      [end of output]
+  
+  note: This error originates from a subprocess, and is likely not a problem with pip.
+error: metadata-generation-failed
+
+× Encountered error while generating package metadata.
+╰─> See above for output.
+
+note: This is an issue with the package mentioned above, not pip.
+hint: See above for details.
+
`],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'elastic' does not start with prefix 'alloy.'"],summaryinfo:[{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/DanielMarchand/aiida-alloy",is_installable:"False"},"aiida-amber":{code_home:"https://github.com/PSDI-UK/aiida-amber",entry_point_prefix:"amber",pip_url:"git+https://github.com/PSDI-UK/aiida-amber",name:"aiida-amber",package_name:"aiida_amber",hosted_on:"github.com",metadata:{description:"AiiDA plugin that wraps amber molecular dynamics executables",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Development Status :: 3 - Alpha","Framework :: AiiDA"],author:"Jas Kalayan",author_email:"jas.kalayan@stfc.ac.uk"},aiida_version:">=2.4.0,<3",entry_points:{"aiida.data":{"amber.sander":"aiida_amber.data.sander:SanderParameters","amber.tleap":"aiida_amber.data.tleap:TleapParameters","amber.tleap_input":"aiida_amber.data.tleap_input:TleapInputData","amber.antechamber":"aiida_amber.data.antechamber:AntechamberParameters","amber.pdb4amber":"aiida_amber.data.pdb4amber:Pdb4amberParameters","amber.parmed":"aiida_amber.data.parmed:ParmedParameters","amber.parmed_input":"aiida_amber.data.parmed_input:ParmedInputData"},"aiida.calculations":{"amber.sander":{description:["AiiDA calculation plugin wrapping the sander executable.",""," AiiDA plugin wrapper for the amber 'sander' command."],spec:{inputs:[{name:"inpcrd",required:!0,valid_types:"SinglefileData",info:"input initial coordinates and (optionally) velocities and periodic box size."},{name:"mdin",required:!0,valid_types:"SinglefileData",info:"input control data for the min/md run."},{name:"parameters",required:!0,valid_types:"SanderParameters",info:"Command line parameters for sander"},{name:"prmtop",required:!0,valid_types:"SinglefileData",info:"input molecular topology, force field, periodic box type, atom and residue names."},{name:"cein",required:!1,valid_types:"SinglefileData, NoneType",info:"input redox state definitions."},{name:"cerestrt",required:!1,valid_types:"SinglefileData, NoneType",info:"redox state definitions, final redox states for restart (same format as cein)."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"cpin",required:!1,valid_types:"SinglefileData, NoneType",info:"input protonation state definitions."},{name:"cprestrt",required:!1,valid_types:"SinglefileData, NoneType",info:"protonation state definitions, final protonation states for restart (same format as cpin)."},{name:"evbin",required:!1,valid_types:"SinglefileData, NoneType",info:"input input for EVB potentials."},{name:"inpdip",required:!1,valid_types:"SinglefileData, NoneType",info:"input polarizable dipole file, when indmeth=3."},{name:"inptraj",required:!1,valid_types:"SinglefileData, NoneType",info:"input coordinate sets in trajectory format, when imin=5 or 6."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"mtmd",required:!1,valid_types:"SinglefileData, NoneType",info:"input (optional) containing list of files and parameters for targeted MD to multiple targets."},{name:"refc",required:!1,valid_types:"SinglefileData, NoneType",info:"input (optional) reference coords for position restraints; also used for targeted MD."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"mdinfo",required:!0,valid_types:"SinglefileData",info:"output latest mdout-format energy info."},{name:"mdout",required:!0,valid_types:"SinglefileData",info:"output user readable state info and diagnostics -o stdout will send output to stdout (to the terminal) instead of to a file."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"ceout",required:!1,valid_types:"SinglefileData",info:"output redox state data saved over trajectory."},{name:"cpout",required:!1,valid_types:"SinglefileData",info:"output protonation state data saved over trajectory."},{name:"mdcrd",required:!1,valid_types:"SinglefileData",info:"output coordinate sets saved over trajectory."},{name:"mden",required:!1,valid_types:"SinglefileData",info:"output extensive energy data over trajectory (not synchronized with mdcrd or mdvel)."},{name:"mdfrc",required:!1,valid_types:"SinglefileData",info:"output force sets saved over trajectory."},{name:"mdvel",required:!1,valid_types:"SinglefileData",info:"output velocity sets saved over trajectory."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"restrt",required:!1,valid_types:"SinglefileData",info:"output final coordinates, velocity, and box dimensions if any - for restarting run."},{name:"rstdip",required:!1,valid_types:"SinglefileData",info:"output polarizable dipole file, when indmeth=3."},{name:"suffix",required:!1,valid_types:"str",info:"output this string will be added to all unspecified output files that are printed (for multisander runs, it will append this suffix to all output files)."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_amber.calculations.sander:SanderCalculation"},"amber.tleap":{description:["AiiDA calculation plugin wrapping the tleap executable.",""," AiiDA plugin wrapper for the amber 'tleap' command."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"TleapParameters",info:"Command line parameters for tleap"},{name:"tleapscript",required:!0,valid_types:"SinglefileData",info:"input file for tleap commands"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"dirs",required:!1,valid_types:"FolderData",info:"path to directory where custom leaprc are"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"tleap_dirs",required:!1,valid_types:"FolderData",info:"path to directory where inputs referenced in tleap input file are"},{name:"tleap_inpfiles",required:!1,valid_types:"SinglefileData",info:"inputs referenced in tleap input file"},{name:"tleap_outfiles",required:!1,valid_types:"List, NoneType",info:"List of tleap output file names."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_amber.calculations.tleap:TleapCalculation"},"amber.antechamber":{description:["AiiDA calculation plugin wrapping the antechamber executable.",""," AiiDA plugin wrapper for the amber 'antechamber' command."],spec:{inputs:[{name:"input_file",required:!0,valid_types:"SinglefileData",info:"input structure file for antechamber"},{name:"parameters",required:!0,valid_types:"AntechamberParameters",info:"Command line parameters for antechamber"},{name:"additional_file",required:!1,valid_types:"SinglefileData, NoneType",info:"additional file"},{name:"antechamber_outfiles",required:!1,valid_types:"List, NoneType",info:"List of antechamber output file names."},{name:"charge_file",required:!1,valid_types:"SinglefileData, NoneType",info:"charge file"},{name:"check_file",required:!1,valid_types:"SinglefileData, NoneType",info:"check file for gaussian"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"esp_file",required:!1,valid_types:"SinglefileData, NoneType",info:"gaussian esp file"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"res_top_file",required:!1,valid_types:"SinglefileData, NoneType",info:"residue toplogy file"}],outputs:[{name:"output_file",required:!0,valid_types:"SinglefileData",info:"output file"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_amber.calculations.antechamber:AntechamberCalculation"},"amber.pdb4amber":{description:["AiiDA calculation plugin wrapping the pdb4amber executable.",""," AiiDA plugin wrapper for the amber 'pdb4amber' command."],spec:{inputs:[{name:"input_file",required:!0,valid_types:"SinglefileData",info:"input pdb file for pdb4amber"},{name:"parameters",required:!0,valid_types:"Pdb4amberParameters",info:"Command line parameters for pdb4amber"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"pdb4amber_outfiles",required:!1,valid_types:"List, NoneType",info:"List of pdb4amber output file names."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"output_file",required:!0,valid_types:"SinglefileData",info:"outputted pdb file"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_amber.calculations.pdb4amber:Pdb4amberCalculation"},"amber.parmed":{description:["AiiDA calculation plugin wrapping the parmed executable.",""," AiiDA plugin wrapper for the amber 'parmed' command."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"ParmedParameters",info:"Command line parameters for parmed"},{name:"parmed_script",required:!0,valid_types:"SinglefileData",info:"input file for parmed commands"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"inpcrd_files",required:!1,valid_types:"SinglefileData",info:"List of inpcrd files to load into ParmEd"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parmed_dirs",required:!1,valid_types:"FolderData",info:"path to directory where inputs referenced in parmed input file are"},{name:"parmed_inpfiles",required:!1,valid_types:"SinglefileData",info:"inputs referenced in parmed input script"},{name:"parmed_outfiles",required:!1,valid_types:"List, NoneType",info:"List of parmed output file names."},{name:"prmtop_files",required:!1,valid_types:"SinglefileData",info:"List of topology files to load into ParmEd"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_amber.calculations.parmed:ParmedCalculation"}},"aiida.parsers":{"amber.sander":"aiida_amber.parsers.sander:SanderParser","amber.tleap":"aiida_amber.parsers.tleap:TleapParser","amber.antechamber":"aiida_amber.parsers.antechamber:AntechamberParser","amber.pdb4amber":"aiida_amber.parsers.pdb4amber:Pdb4amberParser","amber.parmed":"aiida_amber.parsers.parmed:ParmedParser"}},commits_count:115,development_status:"alpha",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:5},{colorclass:"brown",text:"Parsers",count:5},{colorclass:"red",text:"Data",count:7}],pip_install_cmd:"pip install git+https://github.com/PSDI-UK/aiida-amber",is_installable:"True"},"aiida-ase":{code_home:"https://github.com/aiidateam/aiida-ase",documentation_url:"https://aiida-ase.readthedocs.io/",entry_point_prefix:"ase",pip_url:"aiida-ase",plugin_info:"https://raw.github.com/aiidateam/aiida-ase/master/setup.json",name:"aiida-ase",package_name:"aiida_ase",hosted_on:"github.com",metadata:{release_date:"2023-10-04",description:"The official AiiDA plugin for ASE.",author_email:"The AiiDA team ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering"],version:"3.0.0"},aiida_version:">=2.0,<3.0",entry_points:{"aiida.calculations":{"ase.ase":{description:["`CalcJob` implementation that can be used to wrap around the ASE calculators."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters for the namelists."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"The k-points to use for the calculation."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional settings that control the plugin."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"array",required:!1,valid_types:"ArrayData",info:""},{name:"parameters",required:!1,valid_types:"Dict",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"structure",required:!1,valid_types:"StructureData",info:""},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"One of the expected output files was missing."},{status:301,message:"The log file from the DFT code was not written out."},{status:302,message:"Relaxation did not complete."},{status:303,message:"SCF Failed."},{status:305,message:"Cannot identify what went wrong."},{status:306,message:"gpaw could not find the PAW potentials."},{status:307,message:"Attribute Error found in the stderr file."},{status:308,message:"Fermi level is infinite."},{status:400,message:"The calculation ran out of walltime."}]},class:"aiida_ase.calculations.ase:AseCalculation"}},"aiida.parsers":{"ase.ase":"aiida_ase.parsers.ase:AseParser","ase.gpaw":"aiida_ase.parsers.gpaw:GpawParser"},"aiida.workflows":{"ase.gpaw.base":{description:["Workchain to run a GPAW calculation with automated error handling and restarts."],spec:{inputs:[{name:"gpaw",required:!0,valid_types:"Data",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"k-points to use for the calculation."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"array",required:!1,valid_types:"ArrayData",info:""},{name:"parameters",required:!1,valid_types:"Dict",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"structure",required:!1,valid_types:"StructureData",info:""},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_ase.workflows.base:GpawBaseWorkChain"}}},commits_count:1,development_status:"beta",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-ase",is_installable:"True"},"aiida-autocas":{entry_point_prefix:"autocas",code_home:"https://github.com/microsoft/aiida-autocas",version_file:"https://raw.githubusercontent.com/microsoft/aiida-autocas/main/aiida_autocas/__init__.py",pip_url:"git+https://github.com/microsoft/aiida-autocas",name:"aiida-autocas",package_name:"aiida_autocas",hosted_on:"github.com",metadata:{version:"0.1.0",description:"AiiDA AutoCAS Plugin",classifiers:[]},aiida_version:">=2.0,<3",entry_points:{"aiida.calculations":{autocas:"aiida_autocas.calculations:AutoCASCalculation"},"aiida.parsers":{autocas:"aiida_autocas.parsers:AutoCASParser"}},commits_count:0,development_status:"planning",errors:[],warnings:["W003: Missing classifier 'Framework :: AiiDA'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install git+https://github.com/microsoft/aiida-autocas"},"aiida-bader":{entry_point_prefix:"bader",pip_url:"aiida-bader",code_home:"https://github.com/superstar54/aiida-bader",documentation_url:"https://aiida-bader.readthedocs.io/en/latest/",name:"aiida-bader",package_name:"aiida_bader",hosted_on:"github.com",metadata:{release_date:"2024-10-10",description:"Collections of node-based workflow created using aiida-workgraph.",author_email:"Xing Wang ",classifiers:["Development Status :: 1 - Planning","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering"],version:"0.0.11"},aiida_version:"*",entry_points:{"aiida.calculations":{bader:"aiida_bader.calculations:BaderCalculation"},"aiida.parsers":{bader:"aiida_bader.parsers:BaderParser"},"aiida.workflows":{"bader.qe":"aiida_bader.workchains:QeBaderWorkChain"},"aiidalab_qe.properties":{bader:"aiida_bader.qeapp:bader"}},commits_count:17,development_status:"planning",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1},{colorclass:"orange",text:"Other (Aiidalab qe properties)",count:1}],pip_install_cmd:"pip install aiida-bader"},"aiida-bands-inspect":{code_home:"https://github.com/greschd/aiida-bands-inspect",documentation_url:"https://aiida-bands-inspect.readthedocs.io",entry_point_prefix:"bands_inspect",pip_url:"aiida-bands-inspect",name:"aiida-bands-inspect",package_name:"aiida_bands_inspect",hosted_on:"github.com",metadata:{release_date:"2020-03-26",description:"AiiDA Plugin for running bands_inspect",author:"Dominik Gresch",author_email:"greschd@gmx.ch",license:"Apache 2.0",home_page:"https://aiida-bands-inspect.readthedocs.io",classifiers:["Development Status :: 4 - Beta","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: Apache Software License","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Scientific/Engineering :: Physics"],version:"0.4.0"},aiida_version:null,entry_points:{"aiida.calculations":{"bands_inspect.align":"aiida_bands_inspect.calculations.align:AlignCalculation","bands_inspect.difference":"aiida_bands_inspect.calculations.difference:DifferenceCalculation","bands_inspect.plot":"aiida_bands_inspect.calculations.plot:PlotCalculation"},"aiida.parsers":{"bands_inspect.bands":"aiida_bands_inspect.parsers.bands:BandsParser","bands_inspect.difference":"aiida_bands_inspect.parsers.difference:DifferenceParser","bands_inspect.align":"aiida_bands_inspect.parsers.align:AlignParser","bands_inspect.plot":"aiida_bands_inspect.parsers.plot:PlotParser"}},commits_count:0,development_status:"beta",errors:[`E002: Failed to import package aiida_bands_inspect
Traceback (most recent call last):
+  File "", line 1, in 
+  File "/opt/conda/lib/python3.10/site-packages/aiida_bands_inspect/__init__.py", line 8, in 
+    from . import calculations
+  File "/opt/conda/lib/python3.10/site-packages/aiida_bands_inspect/calculations/__init__.py", line 9, in 
+    from . import difference
+  File "/opt/conda/lib/python3.10/site-packages/aiida_bands_inspect/calculations/difference.py", line 11, in 
+    from aiida.orm import JobCalculation, DataFactory
+ImportError: cannot import name 'JobCalculation' from 'aiida.orm' (/opt/conda/lib/python3.10/site-packages/aiida/orm/__init__.py)
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:3},{colorclass:"brown",text:"Parsers",count:4}],pip_install_cmd:"pip install aiida-bands-inspect",is_installable:"True"},"aiida-bigdft":{code_home:"https://github.com/BigDFT-group/aiida-bigdft-plugin",development_status:"beta",entry_point_prefix:"bigdft",pip_url:"aiida-bigdft",plugin_info:"https://raw.github.com/BigDFT-group/aiida-bigdft-plugin/master/setup.json",name:"aiida-bigdft",package_name:"aiida_bigdft",hosted_on:"github.com",metadata:{release_date:"2024-01-10",description:"Translation layer for AiiDA-PyBigDFT",author_email:"Louis Beal ",classifiers:["Development Status :: 3 - Alpha","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.3.0"},aiida_version:">=1.6.3,<3",entry_points:{"aiida.calculations":{bigdft:{description:["AiiDA plugin wrapping a BigDFT calculation",""," requires a valid BigDFT install and a copy of `bigdft.py` on the target machine."],spec:{inputs:[{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure (AiiDA format)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"extra_files_recv",required:!1,valid_types:"List",info:"Extra files to retrieve from calculation"},{name:"extra_files_send",required:!1,valid_types:"List",info:"Extra files to send with calculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"BigDFTParameters",info:"BigDFT Inputfile parameters, as Dict"},{name:"params_fname",required:!1,valid_types:"Str",info:"Name override for parameters file"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"structure_fname",required:!1,valid_types:"Str",info:"Name override for structure file"}],outputs:[{name:"energy",required:!0,valid_types:"Float",info:"Final energy estimate taken from logfile"},{name:"logfile",required:!0,valid_types:"BigDFTLogfile",info:"BigDFT calculation Logfile"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"timefile",required:!0,valid_types:"BigDFTFile",info:"BigDFT calculation time log"},{name:"ttotal",required:!0,valid_types:"Float",info:"Estimated total run time (excluding queue)"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:100,message:"Calculation did not produce all expected output files."},{status:101,message:"Calculation did not produce all expected output files."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:400,message:"Calculation did not finish because of a walltime issue."},{status:401,message:"Calculation did not finish because of memory limit"}]},class:"aiida_bigdft.calculations:BigDFTCalculation"}},"aiida.cmdline.data":{bigdft:"aiida_bigdft.cli:data_cli"},"aiida.data":{bigdft:"aiida_bigdft.data.BigDFTParameters:BigDFTParameters",bigdftfile:"aiida_bigdft.data.BigDFTFile:BigDFTFile",bigdftlogfile:"aiida_bigdft.data.BigDFTFile:BigDFTLogfile"},"aiida.parsers":{bigdft:"aiida_bigdft.parsers:BigDFTParser"},"aiida.workflows":{bigdft:{description:["Base workchain for running a BigDFT Calculation"],spec:{inputs:[{name:"BigDFT",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"energy",required:!0,valid_types:"Float",info:"Final energy estimate taken from logfile"},{name:"logfile",required:!0,valid_types:"BigDFTLogfile",info:"BigDFT calculation Logfile"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"timefile",required:!0,valid_types:"BigDFTFile",info:"BigDFT calculation time log"},{name:"ttotal",required:!0,valid_types:"Float",info:"Estimated total run time (excluding queue)"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation encountered an unrecoverable error"},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_bigdft.workflows.base:BigDFTBaseWorkChain"},"bigdft.relax":"aiida_bigdft.workflows.relax:BigDFTRelaxWorkChain"}},commits_count:21,errors:[],warnings:["W005: Development status in classifiers (alpha) does not match development_status in metadata (beta)","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:3},{colorclass:"green",text:"Workflows",count:2},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install aiida-bigdft",is_installable:"True"},"aiida-castep":{code_home:"https://gitlab.com/bz1/aiida-castep",development_status:"stable",documentation_url:"https://aiida-castep.readthedocs.io/",entry_point_prefix:"castep",pip_url:"aiida-castep",plugin_info:"https://gitlab.com/bz1/aiida-castep/raw/master/setup.json",name:"aiida-castep",package_name:"aiida_castep",hosted_on:"gitlab.com",metadata:{release_date:"2022-05-26",description:"AiiDA plugin for CASTEP",author:"Bonan Zhu",author_email:"zhubonan@outlook.com",license:"MIT License",home_page:"https://github.com/zhubonan/aiida-castep",classifiers:["Framework :: AiiDA","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"2.0.1"},aiida_version:">=2.0,<3.0",entry_points:{"aiida.calculations":{"castep.castep":{description:["Class representing a generic CASTEP calculation -"," This class should work for all types of calculations."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"A node that defines the input parameters"},{name:"pseudos",required:!0,valid_types:"",info:"Use nodes for the pseudopotentails of one ofthe element in the structure. You should pass aa dictionary specifying the pseudpotential node foreach kind such as {O: }"},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure"},{name:"bs_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: bandstructure"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"elnes_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: elnes"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Use a node defining the kpoints for the calculation"},{name:"magres_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: magres"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optics_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: optics"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote folder as the parent folder. Useful for restarts."},{name:"phonon_fine_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: phonon, phonon+efield"},{name:"phonon_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: phonon, phonon+efield"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"A node for additional settings"},{name:"spectral_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: spectral"},{name:"supercell_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: phonon"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"Parsed results in a dictionary format."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:0,message:"Calculation terminated gracefully, end found"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:101,message:"SCF Cycles failed to reach convergence"},{status:103,message:"Stopped execuation due to detection of 'stop ' keyword in param file."},{status:104,message:"CASTEP generate error files. Check them for details"},{status:105,message:"Cannot find the end of calculation"},{status:106,message:"No output .castep files found"},{status:107,message:"Calculation self-terminated due to time limit"},{status:108,message:"No retrieve folder is found"},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"UNKOWN ERROR"},{status:501,message:"At least one kpoints/spin has no empty bands - please rerun with increased nextra_bands."}]},class:"aiida_castep.calculations.castep:CastepCalculation"},"castep.ts":{description:["CASTEP calculation for transition state search. Use an extra input product structure."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"A node that defines the input parameters"},{name:"product_structure",required:!0,valid_types:"StructureData",info:"Product structure for transition state search."},{name:"pseudos",required:!0,valid_types:"",info:"Use nodes for the pseudopotentails of one ofthe element in the structure. You should pass aa dictionary specifying the pseudpotential node foreach kind such as {O: }"},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure"},{name:"bs_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: bandstructure"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"elnes_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: elnes"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Use a node defining the kpoints for the calculation"},{name:"magres_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: magres"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optics_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: optics"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote folder as the parent folder. Useful for restarts."},{name:"phonon_fine_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: phonon, phonon+efield"},{name:"phonon_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: phonon, phonon+efield"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"A node for additional settings"},{name:"spectral_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: spectral"},{name:"supercell_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Extra kpoints input for task: phonon"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"Parsed results in a dictionary format."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:0,message:"Calculation terminated gracefully, end found"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:101,message:"SCF Cycles failed to reach convergence"},{status:103,message:"Stopped execuation due to detection of 'stop ' keyword in param file."},{status:104,message:"CASTEP generate error files. Check them for details"},{status:105,message:"Cannot find the end of calculation"},{status:106,message:"No output .castep files found"},{status:107,message:"Calculation self-terminated due to time limit"},{status:108,message:"No retrieve folder is found"},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"UNKOWN ERROR"},{status:501,message:"At least one kpoints/spin has no empty bands - please rerun with increased nextra_bands."}]},class:"aiida_castep.calculations.castep:CastepTSCalculation"}},"aiida.cmdline.data":{"castep-helper":"aiida_castep.cmdline.helper_cmd:helper_cmd","castep-pseudos":"aiida_castep.cmdline.otfg_cmd:pseudos_cmd"},"aiida.data":{"castep.otfgdata":"aiida_castep.data.otfg:OTFGData","castep.uspdata":"aiida_castep.data.usp:UspData"},"aiida.groups":{"castep.otfg":"aiida_castep.data.otfg:OTFGGroup"},"aiida.parsers":{"castep.castep":"aiida_castep.parsers.castep:CastepParser"},"aiida.tests":{"castep.calculation":"aiida_castep.tests.dbtests.dbtestcalculation"},"aiida.tools.calculations":{"castep.castep":"aiida_castep.calculations.tools:CastepCalcTools"},"aiida.workflows":{"castep.altrelax":{description:["A relaxation workflow that alternates between fixed cell and unfixed cell"," This is meidate the problem in CASTEP where if the cell is partially constraints"," the convergence would be very slow.",""," To overcome this problem, the structure should be relaxed with cell constraints"," then restart with fixed cell and repeat.",""," Following fields can be used in ``relax_options``",""," :var_cell_iter_max: Maximum iterations in variable cell relaxation, default to 10",""," :fix_cell_iter_max: Maximum iterations in fixed cell relaxation, default to 20"],spec:{inputs:[{name:"base",required:!0,valid_types:"Data",info:""},{name:"calc",required:!0,valid_types:"Data",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"Structure to be used for relaxation."},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:"Wether to clean the workdir of the calculations at the end of the workchain. The default is not performing any cleaning."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax_options",required:!1,valid_types:"Dict, NoneType",info:"Options for relaxation."}],outputs:[{name:"output_bands",required:!0,valid_types:"BandsData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:""},{name:"output_array",required:!1,valid_types:"ArrayData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed structure."},{name:"output_trajectory",required:!1,valid_types:"ArrayData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:101,message:"Subprocess lauched has failed in the relax stage"},{status:102,message:"Geometry optimisation is not converged but the maximum iteration is exceeded."},{status:201,message:"NO cell_constraints find in the input"}]},class:"aiida_castep.workflows.relax:CastepAlterRelaxWorkChain"},"castep.bands":{description:["Workchain for running bands calculation.",""," This workchain does the following:",""," 1. Relax the structure if requested (eg. inputs passed to the relax namespace)."," 2. Optionally: Do a SCF singlepoint calculation"," 3. Do combined SCF + non-SCF calculation for bands and dos.",""," Inputs must be passed for the SCF calculation (dispatched to bands and DOS),"," others are optional.",""," Input for bands and dos calculations are optional. However, if they are needed, the full list of inputs must"," be passed. For the `parameters` node, one may choose to only specify those fields that need to be updated."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:"Inputs for SCF workchain, mandatory. Used as template for bands/dos if not supplied separately"},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure"},{name:"bands",required:!1,valid_types:"Data",info:"Inputs for bands calculation, if needed"},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Explicit kpoints for the bands"},{name:"bands_kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"Spacing for band distances, used by seekpath"},{name:"clean_children_workdir",required:!1,valid_types:"Str, NoneType",info:"What part of the called children to clean"},{name:"dos",required:!1,valid_types:"Data",info:"Inputs for DOS calculation, if needed"},{name:"dos_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Kpoints for running DOS calculations"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"only_dos",required:!1,valid_types:"",info:"Flag for running only DOS calculations"},{name:"options",required:!1,valid_types:"",info:"Options for this workchain. Supported keywords: dos_smearing, dos_npoints."},{name:"relax",required:!1,valid_types:"Data",info:"Inputs for Relaxation workchain, if needed"},{name:"run_separate_scf",required:!1,valid_types:"",info:"Flag for running a separate SCF calculation, default to False"}],outputs:[{name:"band_structure",required:!0,valid_types:"",info:"Computed band structure with labels"},{name:"dos_bands",required:!1,valid_types:"",info:"Bands from the DOS calculation"},{name:"primitive_structure",required:!1,valid_types:"",info:"Primitive structure used for band structure calculations"},{name:"seekpath_parameters",required:!1,valid_types:"",info:"Parameters used by seekpath"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:501,message:"Relaxation workchain failed"},{status:502,message:"SCF workchain failed"},{status:503,message:"Band structure workchain failed"},{status:504,message:"DOS workchain failed"}]},class:"aiida_castep.workflows.bands:CastepBandsWorkChain"},"castep.base":{description:["A basic workchain for generic CASTEP calculations."," We try to handle erros such as walltime exceeded or SCF not converged"],spec:{inputs:[{name:"calc",required:!0,valid_types:"Data",info:""},{name:"calc_options",required:!1,valid_types:"Dict, NoneType",info:"Options to be passed to calculations's metadata.options"},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:"Wether to clean the workdir of the calculations or not, the default is not clean."},{name:"continuation_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote folder as the parent folder. Useful for restarts."},{name:"ensure_gamma_centering",required:!1,valid_types:"Bool, NoneType",info:"Ensure the kpoint grid is gamma centred."},{name:"kpoints_spacing",required:!1,valid_types:"Float, NoneType",info:"Kpoint spacing"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of restarts"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Options specific to the workchain.Avaliable options: queue_wallclock_limit, use_castep_bin"},{name:"pseudos_family",required:!1,valid_types:"Str, NoneType",info:"Pseudopotential family to be used"},{name:"reuse_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote folder as the parent folder. Useful for restarts."}],outputs:[{name:"output_bands",required:!0,valid_types:"BandsData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:""},{name:"output_array",required:!1,valid_types:"ArrayData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:""},{name:"output_trajectory",required:!1,valid_types:"ArrayData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"The maximum number of iterations has been exceeded"},{status:201,message:"The maximum length of the wallclocks has been exceeded"},{status:301,message:"CASTEP generated error files and is not recoverable"},{status:302,message:"Cannot reach SCF convergence despite restart efforts"},{status:400,message:"The stop flag has been put in the .param file to request termination of the calculation."},{status:900,message:"Input validate is failed"},{status:901,message:"Completed one iteration but found not calculation returned"},{status:1e3,message:"Error is not known"}]},class:"aiida_castep.workflows.base:CastepBaseWorkChain"},"castep.relax":{description:["WorkChain to relax structures."," Restart the relaxation calculation until the structure is fully relaxed."," Each CASTEP relaxation may finish without error with not fully relaxed structure"," if the number of iteration is exceeded (*geom_max_iter*)."," This workchain try to restart such calculations (wrapped in CastepBaseWorkChain)"," until the structure is fully relaxed",""," ``relax_options`` is a Dict of the options avaliable fields are:",""," - restart_mode: mode of restart, choose from ``reuse`` (default), ``structure``,"," ``continuation``."," - bypass: Bypass relaxation control - e.g. no checking of the convergence."," Can be used for doing singlepoint calculation."],spec:{inputs:[{name:"base",required:!0,valid_types:"Data",info:""},{name:"calc",required:!0,valid_types:"Data",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"Structure to be used for relaxation."},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:"Wether to clean the workdir of the calculations at the end of the workchain. The default is not performing any cleaning."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax_options",required:!1,valid_types:"Dict, NoneType",info:"Options for relaxation."}],outputs:[{name:"output_bands",required:!0,valid_types:"BandsData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:""},{name:"output_array",required:!1,valid_types:"ArrayData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed structure."},{name:"output_trajectory",required:!1,valid_types:"ArrayData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:101,message:"Subprocess lauched has failed in the relax stage"},{status:102,message:"Geometry optimisation is not converged but the maximum iteration is exceeded."}]},class:"aiida_castep.workflows.relax:CastepRelaxWorkChain"}},console_scripts:{"castep.mock":"aiida_castep.cmdline.mock_castep:mock_castep"}},commits_count:0,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:2},{colorclass:"green",text:"Workflows",count:4},{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Data commands, Groups, Tests, ...)",count:5}],pip_install_cmd:"pip install aiida-castep",is_installable:"True"},"aiida-catmap":{code_home:"https://github.com/sudarshanv01/aiida-catmap",entry_point_prefix:"catmap",name:"aiida-catmap",package_name:"aiida_catmap",hosted_on:"github.com",metadata:{author:"Sudarshan Vijay",author_email:"vijays@fysik.dtu.dk",version:"0.2.0a0",description:"AiiDA package that interfaces with Kinetic modelling code CatMAP",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=1.1.0,<2.0.0",entry_points:{"aiida.calculations":{catmap:"aiida_catmap.calculations.catmap:CatMAPCalculation"},"aiida.parsers":{catmap:"aiida_catmap.parsers.catmap:CatMAPParser"}},commits_count:0,development_status:"planning",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"See source code repository."},"aiida-catmat":{code_home:"https://github.com/pzarabadip/aiida-catmat",entry_point_prefix:"catmat",development_status:"beta",documentation_url:"https://aiida-catmat.readthedocs.io/",pip_url:"aiida-catmat",name:"aiida-catmat",package_name:"aiida_catmat",hosted_on:"github.com",metadata:{release_date:"2022-07-21",description:"Collection of AiiDA WorkChains Developed in Morgan Group",author:"Pezhman Zarabadi-Poor",author_email:"pzarabadip@gmail.com",license:"MIT License",home_page:"https://github.com/pzarabadip/aiida-catmat",classifiers:["Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"1.0.0b0"},aiida_version:null,entry_points:{"aiida.parsers":{vasp_base_parser:"aiida_catmat.parsers:VaspBaseParser"},"aiida.workflows":{"vasp.base":"aiida_catmat.workchains:VaspBaseWorkChain","catmat.vasp_multistage":"aiida_catmat.workchains:VaspMultiStageWorkChain","catmat.vasp_converge":"aiida_catmat.workchains:VaspConvergeWorkChain","catmat.vasp_catmat":"aiida_catmat.workchains:VaspCatMatWorkChain","catmat.vasp_multistage_ddec":"aiida_catmat.workchains:VaspMultiStageDdecWorkChain"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-catmat
Collecting aiida-catmat
+  Downloading aiida-catmat-1.0.0b0.tar.gz (27 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+ERROR: Packages installed from PyPI cannot depend on packages which are not also hosted on PyPI.
+aiida-catmat depends on aiida-vasp@ git+https://github.com/aiida-vasp/aiida-vasp.git@14771c14ac5ccb08ac521d5936eb0c4ec5b9337f
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found","W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'vasp_base_parser' does not start with prefix 'catmat.'","W010: Entry point 'vasp.base' does not start with prefix 'catmat.'"],summaryinfo:[{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:5}],pip_install_cmd:"pip install --pre aiida-catmat",is_installable:"False"},"aiida-ce":{code_home:"https://github.com/unkcpz/aiida-ce",development_status:"beta",entry_point_prefix:"ce",pip_url:"git+https://github.com/unkcpz/aiida-ce",name:"aiida-ce",package_name:"aiida_ce",hosted_on:"github.com",metadata:{author:"unkcpz",author_email:"morty.yu@yahoo.com",version:"0.1.0a0",description:"AiiDA plugin for running cluster expansion using icet.",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=1.0.0,<2.0.0",entry_points:{"aiida.data":{ce:"aiida_ce.data:DiffParameters","ce.structures":"aiida_ce.data.structure_set:StructureSet","ce.cluster":"aiida_ce.data.cluster:ClusterSpaceData"},"aiida.calculations":{"ce.genenum":"aiida_ce.calculations.genenum:EnumCalculation","ce.gensqs":"aiida_ce.calculations.gensqs:SqsCalculation","ce.train":"aiida_ce.calculations.train:TrainCalculation"},"aiida.parsers":{"ce.genenum":"aiida_ce.parsers.genenum:EnumParser","ce.gensqs":"aiida_ce.parsers.gensqs:SqsParser","ce.train":"aiida_ce.parsers.train:TrainParser"},"aiida.cmdline.data":{ce:"aiida_ce.cli:data_cli"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-ce
Collecting git+https://github.com/unkcpz/aiida-ce
+  Cloning https://github.com/unkcpz/aiida-ce to /tmp/pip-req-build-dr7m931n
+  Running command git clone --filter=blob:none --quiet https://github.com/unkcpz/aiida-ce /tmp/pip-req-build-dr7m931n
+  Resolved https://github.com/unkcpz/aiida-ce to commit 84cbc7c1e14af3cac5530e092e529bc84cfcfe6d
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Installing backend dependencies: started
+  Installing backend dependencies: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-ce to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-ce==0.1.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-ce 0.1.0 depends on aiida-core<2.0.0 and >=1.5.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:3},{colorclass:"brown",text:"Parsers",count:3},{colorclass:"red",text:"Data",count:3},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install git+https://github.com/unkcpz/aiida-ce",is_installable:"False"},"aiida-champ":{code_home:"https://github.com/TREX-CoE/aiida-champ",development_status:"beta",documentation_url:"http://aiida-champ.readthedocs.io/",entry_point_prefix:"champ",pip_url:"aiida-champ",name:"aiida-champ",package_name:"aiida_champ",hosted_on:"github.com",metadata:{release_date:"2021-12-27",description:"AiiDA plugin that wraps the vmc executable of CHAMP code for computing the total energy and much more stuff.",author:"Ravindra Shinde",author_email:"r.l.shinde@utwente.nl",license:"MIT",home_page:"https://github.com/neelravi/aiida-champ",classifiers:["Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"1.2.6"},aiida_version:null,entry_points:{"aiida.data":{CHAMP:"aiida_champ.data:CHAMPParameters"},"aiida.calculations":{CHAMP:"aiida_champ.calculations:CHAMPCalculation"},"aiida.parsers":{CHAMP:"aiida_champ.parsers:CHAMPParser"},"aiida.cmdline.data":{CHAMP:"aiida_champ.cli:data_cli"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-champ
Collecting aiida-champ
+  Downloading aiida-champ-1.2.6.tar.gz (10 kB)
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Installing backend dependencies: started
+  Installing backend dependencies: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-champ to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-champ==1.2.6 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-champ 1.2.6 depends on aiida-core<2.0.0 and >=1.1.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'CHAMP' does not start with prefix 'champ.'","W010: Entry point 'CHAMP' does not start with prefix 'champ.'","W010: Entry point 'CHAMP' does not start with prefix 'champ.'","W010: Entry point 'CHAMP' does not start with prefix 'champ.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install aiida-champ",is_installable:"False"},"aiida-codtools":{code_home:"https://github.com/aiidateam/aiida-codtools",documentation_url:"https://aiida-codtools.readthedocs.io/",entry_point_prefix:"codtools",pip_url:"aiida-codtools",plugin_info:"https://raw.githubusercontent.com/aiidateam/aiida-codtools/master/setup.json",name:"aiida-codtools",package_name:"aiida_codtools",hosted_on:"github.com",metadata:{release_date:"2023-02-02",description:"The Official AiiDA plugin for the cod-tools package.",author_email:"The AiiDA team ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"3.1.0"},aiida_version:">=2.1,<3.0",entry_points:{"aiida.calculations":{"codtools.cif_base":{description:["Generic `CalcJob` implementation that can easily be extended to work with any of the `cod-tools` scripts."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_base:CifBaseCalculation"},"codtools.cif_cell_contents":{description:["CalcJob plugin for the `cif_cell_contents` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"formulae",required:!0,valid_types:"Dict",info:"A dictionary of formulae present in the CIF."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_cell_contents:CifCellContentsCalculation"},"codtools.cif_cod_check":{description:["CalcJob plugin for the `cif_cod_check` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"messages",required:!0,valid_types:"Dict",info:"Warning and error messages returned by the script."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_cod_check:CifCodCheckCalculation"},"codtools.cif_cod_deposit":{description:["CalcJob plugin for the `cif_cod_deposit` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:300,message:"The deposition failed for unknown reasons."},{status:310,message:"The deposition failed because the input was invalid."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."},{status:410,message:"The deposition failed because one or more CIFs already exist in the COD."},{status:420,message:"The structure is unchanged and so deposition is unnecessary."}]},class:"aiida_codtools.calculations.cif_cod_deposit:CifCodDepositCalculation"},"codtools.cif_cod_numbers":{description:["CalcJob plugin for the `cif_cod_numbers` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"numbers",required:!0,valid_types:"Dict",info:"Mapping of COD IDs found with their formula and count."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_cod_numbers:CifCodNumbersCalculation"},"codtools.cif_filter":{description:["CalcJob plugin for the `cif_filter` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF produced by the script."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_filter:CifFilterCalculation"},"codtools.cif_select":{description:["CalcJob plugin for the `cif_select` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF produced by the script."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_select:CifSelectCalculation"},"codtools.cif_split_primitive":{description:["CalcJob plugin for the `cif_split_primitive` script of the `cod-tools` package."],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CIF to be processed."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Command line parameters."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"cifs",required:!0,valid_types:"CifData",info:"The CIFs produced by the script."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"messages",required:!1,valid_types:"Dict",info:"Warning and error messages returned by script."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Neither the output for the error file could be read from the retrieved folder."},{status:311,message:"The output file could not be read from the retrieved folder."},{status:312,message:"The error file could not be read from the retrieved folder."},{status:313,message:"The output file is empty."},{status:320,message:"Invalid command line option passed."},{status:400,message:"The output file could not be parsed."},{status:410,message:"The output file could not be parsed into a CifData object."}]},class:"aiida_codtools.calculations.cif_split_primitive:CifSplitPrimitiveCalculation"},"codtools.primitive_structure_from_cif":{description:["Attempt to parse the given `CifData` and create a `StructureData` from it.",""," First the raw CIF file is parsed with the given `parse_engine`. The resulting `StructureData` is then passed through"," SeeKpath to try and get the primitive cell. If that is successful, important structural parameters as determined by"," SeeKpath will be set as extras on the structure node which is then returned as output.",""," :param cif: the `CifData` node"," :param parse_engine: the parsing engine, supported libraries 'ase' and 'pymatgen'"," :param symprec: a `Float` node with symmetry precision for determining primitive cell in SeeKpath"," :param site_tolerance: a `Float` node with the fractional coordinate distance tolerance for finding overlapping"," sites. This will only be used if the parse_engine is pymatgen"," :return: the primitive `StructureData` as determined by SeeKpath"],spec:{inputs:[{name:"cif",required:!0,valid_types:"Data",info:"the `CifData` node"},{name:"parse_engine",required:!0,valid_types:"Data",info:"the parsing engine, supported libraries 'ase' and 'pymatgen'"},{name:"site_tolerance",required:!0,valid_types:"Data",info:"a `Float` node with the fractional coordinate distance tolerance for finding overlapping\nsites. This will only be used if the parse_engine is pymatgen"},{name:"symprec",required:!0,valid_types:"Data",info:"a `Float` node with symmetry precision for determining primitive cell in SeeKpath"},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_codtools.calculations.functions.primitive_structure_from_cif:primitive_structure_from_cif"}},"aiida.parsers":{"codtools.cif_base":"aiida_codtools.parsers.cif_base:CifBaseParser","codtools.cif_cell_contents":"aiida_codtools.parsers.cif_cell_contents:CifCellContentsParser","codtools.cif_cod_check":"aiida_codtools.parsers.cif_cod_check:CifCodCheckParser","codtools.cif_cod_deposit":"aiida_codtools.parsers.cif_cod_deposit:CifCodDepositParser","codtools.cif_cod_numbers":"aiida_codtools.parsers.cif_cod_numbers:CifCodNumbersParser","codtools.cif_split_primitive":"aiida_codtools.parsers.cif_split_primitive:CifSplitPrimitiveParser"},"aiida.workflows":{"codtools.cif_clean":{description:["WorkChain to clean a `CifData` node using the `cif_filter` and `cif_select` scripts of `cod-tools`.",""," It will first run `cif_filter` to correct syntax errors, followed by `cif_select` which will canonicalize the tags."," If a group is passed for the `group_structure` input, the atomic structure library defined by the `engine` input"," will be used to parse the final cleaned `CifData` to construct a `StructureData` object, which will then be passed"," to the `SeeKpath` library to analyze it and return the primitive structure"],spec:{inputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The CifData node that is to be cleaned."},{name:"cif_filter",required:!0,valid_types:"Data",info:""},{name:"cif_select",required:!0,valid_types:"Data",info:""},{name:"group_cif",required:!1,valid_types:"Group, NoneType",info:"An optional Group to which the final cleaned CifData node will be added."},{name:"group_structure",required:!1,valid_types:"Group, NoneType",info:"An optional Group to which the final reduced StructureData node will be added."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parse_engine",required:!1,valid_types:"Str",info:"The atomic structure engine to parse the cif and create the structure."},{name:"site_tolerance",required:!1,valid_types:"Float",info:"The fractional coordinate distance tolerance for finding overlapping sites (pymatgen only)."},{name:"symprec",required:!1,valid_types:"Float",info:"The symmetry precision used by SeeKpath for crystal symmetry refinement."}],outputs:[{name:"cif",required:!0,valid_types:"CifData",info:"The cleaned CifData node."},{name:"structure",required:!1,valid_types:"StructureData",info:"The primitive cell structure created with SeeKpath from the cleaned CifData."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"The CifFilterCalculation step failed."},{status:402,message:"The CifSelectCalculation step failed."},{status:410,message:"The cleaned CifData contains sites with unknown species."},{status:411,message:"The cleaned CifData defines no atomic sites."},{status:412,message:"The cleaned CifData defines sites with attached hydrogens with incomplete positional data."},{status:413,message:"The cleaned CifData defines sites with invalid atomic occupancies."},{status:414,message:"Failed to parse a StructureData from the cleaned CifData."},{status:420,message:"SeeKpath failed to determine the primitive structure."},{status:421,message:"SeeKpath detected inconsistent symmetry operations."}]},class:"aiida_codtools.workflows.cif_clean:CifCleanWorkChain"}},console_scripts:{"aiida-codtools":"aiida_codtools.cli:cmd_root"}},commits_count:0,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:9},{colorclass:"brown",text:"Parsers",count:6},{colorclass:"green",text:"Workflows",count:1},{colorclass:"purple",text:"Console scripts",count:1}],pip_install_cmd:"pip install aiida-codtools",is_installable:"True"},"aiida-core":{code_home:"https://github.com/aiidateam/aiida-core",development_status:"stable",documentation_url:"https://aiida-core.readthedocs.io/",entry_point_prefix:"",package_name:"aiida",pip_url:"aiida-core",plugin_info:"https://raw.githubusercontent.com/aiidateam/aiida-core/master/setup.json",name:"aiida-core",hosted_on:"github.com",metadata:{release_date:"2024-11-06",description:"AiiDA is a workflow manager for computational science with a strong focus on provenance, performance and extensibility.",author_email:"The AiiDA team ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering"],version:"2.6.3"},aiida_version:"==2.6.3",entry_points:{"aiida.brokers":{"core.rabbitmq":"aiida.brokers.rabbitmq.broker:RabbitmqBroker"},"aiida.calculations":{"core.arithmetic.add":{description:["`CalcJob` implementation to add two numbers using bash for testing and demonstration purposes."],spec:{inputs:[{name:"x",required:!0,valid_types:"Int, Float",info:"The left operand."},{name:"y",required:!0,valid_types:"Int, Float",info:"The right operand."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"sum",required:!0,valid_types:"Int, Float",info:"The sum of the left and right operand."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:310,message:"The output file could not be read."},{status:320,message:"The output file contains invalid output."},{status:410,message:"The sum of the operands is a negative number."}]},class:"aiida.calculations.arithmetic.add:ArithmeticAddCalculation"},"core.templatereplacer":{description:["Simple stub of a plugin that can be used to replace some text in a given template."," Can be used for many different codes, or as a starting point to develop a new plugin.",""," This simple plugin takes two node inputs, both of type Dict, with the labels"," 'parameters' and 'template'",""," You can also add other SinglefileData nodes as input, that will be copied according to"," what is written in 'template' (see below).",""," * parameters: a set of parameters that will be used for substitution.",""," * template: can contain the following parameters:",""," * input_file_template: a string with substitutions to be managed with the format()"," function of python, i.e. if you want to substitute a variable called 'varname', you write"," {varname} in the text. See http://www.python.org/dev/peps/pep-3101/ for more"," details. The replaced file will be the input file.",""," * input_file_name: a string with the file name for the input. If it is not provided, no"," file will be created.",""," * output_file_name: a string with the file name for the output. If it is not provided, no"," redirection will be done and the output will go in the scheduler output file.",""," * cmdline_params: a list of strings, to be passed as command line parameters."," Each one is substituted with the same rule of input_file_template. Optional",""," * input_through_stdin: if True, the input file name is passed via stdin. Default is False if missing.",""," * files_to_copy: if defined, a list of tuple pairs, with format ('link_name', 'dest_rel_path');"," for each tuple, an input link to this calculation is looked for, with link labeled 'link_label',"," and with file type 'Singlefile', and the content is copied to a remote file named 'dest_rel_path'"," Errors are raised in the input links are non-existent, or of the wrong type, or if there are"," unused input files.",""," * retrieve_temporary_files: a list of relative filepaths, that if defined, will be retrieved and"," temporarily stored in an unstored FolderData node that will be available during the"," Parser.parser_with_retrieved call under the key specified by the Parser.retrieved_temporary_folder key"],spec:{inputs:[{name:"template",required:!0,valid_types:"Dict",info:"A template for the input file."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"files",required:!1,valid_types:"RemoteData, SinglefileData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters used to replace placeholders in the template."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The temporary retrieved folder data node could not be accessed."},{status:305,message:"The `template` input node did not specify the key `output_file_name`."},{status:310,message:"The output file could not be read from the retrieved folder."},{status:311,message:"A temporary retrieved file could not be read from the temporary retrieved folder."},{status:320,message:"The output file contains invalid output."}]},class:"aiida.calculations.templatereplacer:TemplatereplacerCalculation"},"core.transfer":{description:["Utility to copy files from different FolderData and RemoteData nodes into a single place.",""," The final destination for these files can be either the local repository (by creating a"," new FolderData node to store them) or in the remote computer (by leaving the files in a"," new remote folder saved in a RemoteData node).",""," Only files from the local computer and from remote folders in the same external computer"," can be moved at the same time with a single instance of this CalcJob.",""," The user needs to provide three inputs:",""," * ``instructions``: a dict node specifying which files to copy from which nodes."," * ``source_nodes``: a dict of nodes, each with a unique identifier label as its key."," * ``metadata.computer``: the computer that contains the remote files and will contain"," the final RemoteData node.",""," The ``instructions`` dict must have the ``retrieve_files`` flag. The CalcJob will create a"," new folder in the remote machine (``RemoteData``) and put all the files there and will either:",""," (1) leave them there (``retrieve_files = False``) or ..."," (2) retrieve all the files and store them locally in a ``FolderData`` (``retrieve_files = True``)",""," The `instructions` dict must also contain at least one list with specifications of which files"," to copy and from where. All these lists take tuples of 3 that have the following format:",""," .. code-block:: python",""," ( source_node_key, path_to_file_in_source, path_to_file_in_target)",""," where the ``source_node_key`` has to be the respective one used when providing the node in the"," ``source_nodes`` input nodes dictionary.","",""," The two main lists to include are ``local_files`` (for files to be taken from FolderData nodes)"," and ``remote_files`` (for files to be taken from RemoteData nodes). Alternatively, files inside"," of RemoteData nodes can instead be put in the ``symlink_files`` list: the only difference is that"," files from the first list will be fully copied in the target RemoteData folder, whereas for the"," files in second list only a symlink to the original file will be created there. This will only"," affect the content of the final RemoteData target folder, but in both cases the full file will"," be copied back in the local target FolderData (if ``retrieve_files = True``)."],spec:{inputs:[{name:"instructions",required:!0,valid_types:"Dict",info:"A dictionary containing the `retrieve_files` flag and at least one of the file lists:`local_files`, `remote_files` and/or `symlink_files`."},{name:"source_nodes",required:!0,valid_types:"FolderData, RemoteData",info:"All the nodes that contain files referenced in the instructions."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida.calculations.transfer:TransferCalculation"}},"aiida.calculations.importers":{"core.arithmetic.add":"aiida.calculations.importers.arithmetic.add:ArithmeticAddCalculationImporter"},"aiida.calculations.monitors":{"core.always_kill":"aiida.calculations.monitors.base:always_kill"},"aiida.cmdline.computer.configure":{"core.local":"aiida.transports.plugins.local:CONFIGURE_LOCAL_CMD","core.ssh":"aiida.transports.plugins.ssh:CONFIGURE_SSH_CMD"},"aiida.cmdline.data":{"core.array":"aiida.cmdline.commands.cmd_data.cmd_array:array","core.bands":"aiida.cmdline.commands.cmd_data.cmd_bands:bands","core.cif":"aiida.cmdline.commands.cmd_data.cmd_cif:cif","core.dict":"aiida.cmdline.commands.cmd_data.cmd_dict:dictionary","core.remote":"aiida.cmdline.commands.cmd_data.cmd_remote:remote","core.singlefile":"aiida.cmdline.commands.cmd_data.cmd_singlefile:singlefile","core.structure":"aiida.cmdline.commands.cmd_data.cmd_structure:structure","core.trajectory":"aiida.cmdline.commands.cmd_data.cmd_trajectory:trajectory","core.upf":"aiida.cmdline.commands.cmd_data.cmd_upf:upf"},"aiida.cmdline.data.structure.import":{},"aiida.data":{"core.array":"aiida.orm.nodes.data.array.array:ArrayData","core.array.bands":"aiida.orm.nodes.data.array.bands:BandsData","core.array.kpoints":"aiida.orm.nodes.data.array.kpoints:KpointsData","core.array.projection":"aiida.orm.nodes.data.array.projection:ProjectionData","core.array.trajectory":"aiida.orm.nodes.data.array.trajectory:TrajectoryData","core.array.xy":"aiida.orm.nodes.data.array.xy:XyData","core.base":"aiida.orm.nodes.data:BaseType","core.bool":"aiida.orm.nodes.data.bool:Bool","core.cif":"aiida.orm.nodes.data.cif:CifData","core.code":"aiida.orm.nodes.data.code.legacy:Code","core.code.containerized":"aiida.orm.nodes.data.code.containerized:ContainerizedCode","core.code.installed":"aiida.orm.nodes.data.code.installed:InstalledCode","core.code.portable":"aiida.orm.nodes.data.code.portable:PortableCode","core.dict":"aiida.orm.nodes.data.dict:Dict","core.enum":"aiida.orm.nodes.data.enum:EnumData","core.float":"aiida.orm.nodes.data.float:Float","core.folder":"aiida.orm.nodes.data.folder:FolderData","core.int":"aiida.orm.nodes.data.int:Int","core.jsonable":"aiida.orm.nodes.data.jsonable:JsonableData","core.list":"aiida.orm.nodes.data.list:List","core.numeric":"aiida.orm.nodes.data.numeric:NumericType","core.orbital":"aiida.orm.nodes.data.orbital:OrbitalData","core.remote":"aiida.orm.nodes.data.remote.base:RemoteData","core.remote.stash":"aiida.orm.nodes.data.remote.stash.base:RemoteStashData","core.remote.stash.folder":"aiida.orm.nodes.data.remote.stash.folder:RemoteStashFolderData","core.singlefile":"aiida.orm.nodes.data.singlefile:SinglefileData","core.str":"aiida.orm.nodes.data.str:Str","core.structure":"aiida.orm.nodes.data.structure:StructureData","core.upf":"aiida.orm.nodes.data.upf:UpfData"},"aiida.groups":{core:"aiida.orm.groups:Group","core.auto":"aiida.orm.groups:AutoGroup","core.import":"aiida.orm.groups:ImportGroup","core.upf":"aiida.orm.groups:UpfFamily"},"aiida.node":{data:"aiida.orm.nodes.data.data:Data",process:"aiida.orm.nodes.process.process:ProcessNode","process.calculation":"aiida.orm.nodes.process.calculation.calculation:CalculationNode","process.calculation.calcfunction":"aiida.orm.nodes.process.calculation.calcfunction:CalcFunctionNode","process.calculation.calcjob":"aiida.orm.nodes.process.calculation.calcjob:CalcJobNode","process.workflow":"aiida.orm.nodes.process.workflow.workflow:WorkflowNode","process.workflow.workchain":"aiida.orm.nodes.process.workflow.workchain:WorkChainNode","process.workflow.workfunction":"aiida.orm.nodes.process.workflow.workfunction:WorkFunctionNode"},"aiida.parsers":{"core.arithmetic.add":"aiida.parsers.plugins.arithmetic.add:ArithmeticAddParser","core.templatereplacer":"aiida.parsers.plugins.templatereplacer.parser:TemplatereplacerParser"},"aiida.schedulers":{"core.direct":"aiida.schedulers.plugins.direct:DirectScheduler","core.lsf":"aiida.schedulers.plugins.lsf:LsfScheduler","core.pbspro":"aiida.schedulers.plugins.pbspro:PbsproScheduler","core.sge":"aiida.schedulers.plugins.sge:SgeScheduler","core.slurm":"aiida.schedulers.plugins.slurm:SlurmScheduler","core.torque":"aiida.schedulers.plugins.torque:TorqueScheduler"},"aiida.storage":{"core.psql_dos":"aiida.storage.psql_dos.backend:PsqlDosBackend","core.sqlite_dos":"aiida.storage.sqlite_dos.backend:SqliteDosStorage","core.sqlite_temp":"aiida.storage.sqlite_temp.backend:SqliteTempBackend","core.sqlite_zip":"aiida.storage.sqlite_zip.backend:SqliteZipBackend"},"aiida.tools.calculations":{},"aiida.tools.data.orbitals":{"core.orbital":"aiida.tools.data.orbital.orbital:Orbital","core.realhydrogen":"aiida.tools.data.orbital.realhydrogen:RealhydrogenOrbital"},"aiida.tools.dbexporters":{},"aiida.tools.dbimporters":{"core.cod":"aiida.tools.dbimporters.plugins.cod:CodDbImporter","core.icsd":"aiida.tools.dbimporters.plugins.icsd:IcsdDbImporter","core.materialsproject":"aiida.tools.dbimporters.plugins.materialsproject:MaterialsProjectImporter","core.mpds":"aiida.tools.dbimporters.plugins.mpds:MpdsDbImporter","core.mpod":"aiida.tools.dbimporters.plugins.mpod:MpodDbImporter","core.nninc":"aiida.tools.dbimporters.plugins.nninc:NnincDbImporter","core.oqmd":"aiida.tools.dbimporters.plugins.oqmd:OqmdDbImporter","core.pcod":"aiida.tools.dbimporters.plugins.pcod:PcodDbImporter","core.tcod":"aiida.tools.dbimporters.plugins.tcod:TcodDbImporter"},"aiida.transports":{"core.local":"aiida.transports.plugins.local:LocalTransport","core.ssh":"aiida.transports.plugins.ssh:SshTransport"},"aiida.workflows":{"core.arithmetic.add_multiply":{description:["Add two numbers and multiply it with a third."],spec:{inputs:[{name:"x",required:!0,valid_types:"Data",info:""},{name:"y",required:!0,valid_types:"Data",info:""},{name:"z",required:!0,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida.workflows.arithmetic.add_multiply:add_multiply"},"core.arithmetic.multiply_add":{description:["WorkChain to multiply two numbers and add a third, for testing and demonstration purposes."],spec:{inputs:[{name:"code",required:!0,valid_types:"AbstractCode",info:""},{name:"x",required:!0,valid_types:"Int",info:""},{name:"y",required:!0,valid_types:"Int",info:""},{name:"z",required:!0,valid_types:"Int",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"result",required:!0,valid_types:"Int",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:400,message:"The result is a negative number."}]},class:"aiida.workflows.arithmetic.multiply_add:MultiplyAddWorkChain"}},console_scripts:{runaiida:"aiida.cmdline.commands.cmd_run:run",verdi:"aiida.cmdline.commands.cmd_verdi:verdi"}},commits_count:277,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:3},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:29},{colorclass:"green",text:"Workflows",count:2},{colorclass:"purple",text:"Console scripts",count:2},{colorclass:"orange",text:"Other (Brokers, Calculations importers, Calculations monitors, ...)",count:49}],pip_install_cmd:"pip install aiida-core",is_installable:"True"},"aiida-cp2k":{code_home:"https://github.com/cp2k/aiida-cp2k",entry_point_prefix:"cp2k",pip_url:"aiida-cp2k",plugin_info:"https://raw.githubusercontent.com/cp2k/aiida-cp2k/master/setup.json",name:"aiida-cp2k",package_name:"aiida_cp2k",hosted_on:"github.com",metadata:{release_date:"2023-03-06",description:"The official AiiDA plugin for CP2K.",author:"The AiiDA team",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python :: 3"],version:"2.0.0"},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.calculations":{cp2k:{description:["This is a Cp2kCalculation, subclass of JobCalculation, to prepare input for an ab-initio CP2K calculation.",""," For information on CP2K, refer to: https://www.cp2k.org."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"The input parameters."},{name:"basissets",required:!1,valid_types:"",info:"A dictionary of basissets to be used in the calculations: key is the atomic symbol, value is either a single basisset or a list of basissets. If multiple basissets for a single symbol are passed, it is mandatory to specify a KIND section with a BASIS_SET keyword matching the names (or aliases) of the basissets."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"file",required:!1,valid_types:"SinglefileData, StructureData",info:"Additional input files."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoint mesh."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Working directory of a previously ran calculation to restart from."},{name:"pseudos",required:!1,valid_types:"",info:"A dictionary of pseudopotentials to be used in the calculations: key is the atomic symbol, value is either a single pseudopotential or a list of pseudopotentials. If multiple pseudos for a single symbol are passed, it is mandatory to specify a KIND section with a PSEUDOPOTENTIAL keyword matching the names (or aliases) of the pseudopotentials."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional input parameters."},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:"The main input structure."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The output dictionary containing results of the calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_bands",required:!1,valid_types:"BandsData",info:"Computed electronic band structure."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"The retrieved folder data node could not be accessed."},{status:210,message:"The retrieved folder did not contain the required output file."},{status:301,message:"The output file could not be read."},{status:302,message:"The output file could not be parsed."},{status:303,message:"The output file was incomplete."},{status:304,message:'The output file contains the word "ABORT".'},{status:312,message:"The output structure could not be parsed."},{status:350,message:"The parser raised an unexpected exception."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:500,message:"The ionic minimization cycle did not converge for the given thresholds."},{status:501,message:"The maximum number of optimization steps reached."}]},class:"aiida_cp2k.calculations:Cp2kCalculation"}},"aiida.parsers":{cp2k_advanced_parser:"aiida_cp2k.parsers:Cp2kAdvancedParser",cp2k_base_parser:"aiida_cp2k.parsers:Cp2kBaseParser",cp2k_tools_parser:"aiida_cp2k.parsers:Cp2kToolsParser"},"aiida.workflows":{"cp2k.base":{description:["Workchain to run a CP2K calculation with automated error handling and restarts."],spec:{inputs:[{name:"cp2k",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The output dictionary containing results of the calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"final_input_parameters",required:!1,valid_types:"Dict",info:"The input parameters used for the final calculation."},{name:"output_bands",required:!1,valid_types:"BandsData",info:"Computed electronic band structure."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed with an unidentified unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:310,message:"The calculation failed with a known unrecoverable error."},{status:400,message:"The calculation didn't produce any data to restart from."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_cp2k.workchains:Cp2kBaseWorkChain"}}},commits_count:15,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:3},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-cp2k",is_installable:"True"},"aiida-crystal-dft":{code_home:"https://github.com/tilde-lab/aiida-crystal-dft",development_status:"beta",documentation_url:"https://github.com/tilde-lab/aiida-crystal-dft",entry_point_prefix:"crystal_dft",pip_url:"git+https://github.com/tilde-lab/aiida-crystal-dft",name:"aiida-crystal-dft",package_name:"aiida_crystal_dft",hosted_on:"github.com",metadata:{description:`Yet another AiiDA plugin for CRYSTAL code, mainly intended for use with the cloud infrastructures +(currently, MPDS)`,classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Intended Audience :: Science/Research","Operating System :: OS Independent","Programming Language :: Python","Programming Language :: Python :: 3","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Programming Language :: Python :: 3.10","Topic :: Scientific/Engineering","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics","Topic :: Scientific/Engineering :: Information Analysis"],author:"Andrey Sobolev, based on aiida-crystal17 plugin by Chris Sewell",author_email:"as@tilde.pro"},aiida_version:">=2.0.2",entry_points:{"aiida.data":{"crystal_dft.basis":"aiida_crystal_dft.data.basis:CrystalBasisData","crystal_dft.basis_family":"aiida_crystal_dft.data.basis_family:CrystalBasisFamilyData"},"aiida.calculations":{"crystal_dft.serial":{description:["No description available"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"basis",required:!1,valid_types:"CrystalBasisData",info:""},{name:"basis_family",required:!1,valid_types:"CrystalBasisFamilyData, NoneType",info:""},{name:"guess_oxistates",required:!1,valid_types:"Bool, NoneType",info:""},{name:"high_spin_preferred",required:!1,valid_types:"Bool, NoneType",info:""},{name:"is_magnetic",required:!1,valid_types:"Bool, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"spinlock_steps",required:!1,valid_types:"Int, NoneType",info:""},{name:"use_oxistates",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"oxidation_states",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:""},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"output_wavefunction",required:!1,valid_types:"SinglefileData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"SCF calculation not converged"},{status:301,message:"Geometry optimization failed"},{status:302,message:"Unit cell not neutral"},{status:303,message:"Basis set linearly dependent"},{status:304,message:"Neighbour list too large"},{status:305,message:"No G-vectors left"},{status:306,message:"Collapsed geometry"},{status:307,message:"Closed shell run - spin polarization not allowed"},{status:308,message:"Parameters for model hessian not defined"},{status:309,message:"Fermi energy not in interval"},{status:310,message:"Insufficient indices for Madelung sums"},{status:350,message:"Internal memory error"},{status:360,message:"Inadequate elastic calculation: additional optimization needed"},{status:400,message:"Unknown error"},{status:401,message:"The retrieved folder data node could not be accessed"}]},class:"aiida_crystal_dft.calculations.serial:CrystalSerialCalculation"},"crystal_dft.parallel":{description:["No description available"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"basis",required:!1,valid_types:"CrystalBasisData",info:""},{name:"basis_family",required:!1,valid_types:"CrystalBasisFamilyData, NoneType",info:""},{name:"guess_oxistates",required:!1,valid_types:"Bool, NoneType",info:""},{name:"high_spin_preferred",required:!1,valid_types:"Bool, NoneType",info:""},{name:"is_magnetic",required:!1,valid_types:"Bool, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"spinlock_steps",required:!1,valid_types:"Int, NoneType",info:""},{name:"use_oxistates",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"oxidation_states",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:""},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"output_wavefunction",required:!1,valid_types:"SinglefileData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"SCF calculation not converged"},{status:301,message:"Geometry optimization failed"},{status:302,message:"Unit cell not neutral"},{status:303,message:"Basis set linearly dependent"},{status:304,message:"Neighbour list too large"},{status:305,message:"No G-vectors left"},{status:306,message:"Collapsed geometry"},{status:307,message:"Closed shell run - spin polarization not allowed"},{status:308,message:"Parameters for model hessian not defined"},{status:309,message:"Fermi energy not in interval"},{status:310,message:"Insufficient indices for Madelung sums"},{status:350,message:"Internal memory error"},{status:360,message:"Inadequate elastic calculation: additional optimization needed"},{status:400,message:"Unknown error"},{status:401,message:"The retrieved folder data node could not be accessed"}]},class:"aiida_crystal_dft.calculations.parallel:CrystalParallelCalculation"},"crystal_dft.properties":{description:["AiiDA calculation plugin wrapping the properties executable."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"wavefunction",required:!0,valid_types:"SinglefileData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:""}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_bands",required:!1,valid_types:"BandsData",info:""},{name:"output_bands_down",required:!1,valid_types:"BandsData",info:""},{name:"output_dos",required:!1,valid_types:"ArrayData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The retrieved folder data node could not be accessed"},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_crystal_dft.calculations.properties:PropertiesCalculation"}},"aiida.parsers":{crystal_dft:"aiida_crystal_dft.parsers.cry_pycrystal:CrystalParser","crystal_dft.properties":"aiida_crystal_dft.parsers.properties:PropertiesParser"},"aiida.workflows":{"crystal_dft.base":{description:["Run CRYSTAL calculation"],spec:{inputs:[{name:"basis_family",required:!0,valid_types:"CrystalBasisFamilyData",info:""},{name:"code",required:!0,valid_types:"Code",info:""},{name:"options",required:!0,valid_types:"Dict",info:"Calculation options"},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"restart_params",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_parameters",required:!1,valid_types:"Dict",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:""},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"output_wavefunction",required:!1,valid_types:"SinglefileData",info:""},{name:"oxidation_states",required:!1,valid_types:"Dict",info:""},{name:"primitive_structure",required:!1,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"CRYSTAL error"},{status:400,message:"Unknown error"}]},class:"aiida_crystal_dft.workflows.base:BaseCrystalWorkChain"}},"aiida.cmdline.data":{crystal_dft:"aiida_crystal_dft.cli.basis:basis_set"}},commits_count:12,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:3},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:2},{colorclass:"green",text:"Workflows",count:1},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install git+https://github.com/tilde-lab/aiida-crystal-dft",is_installable:"True"},"aiida-crystal17":{code_home:"https://github.com/aiidaplugins/aiida-crystal17",development_status:"beta",documentation_url:"https://aiida-crystal17.readthedocs.io",entry_point_prefix:"crystal17",pip_url:"aiida-crystal17",plugin_info:"https://raw.githubusercontent.com/aiidaplugins/aiida-crystal17/master/setup.json",name:"aiida-crystal17",package_name:"aiida_crystal17",hosted_on:"github.com",metadata:{release_date:"2020-09-29",description:"AiiDA plugin for running the CRYSTAL17 code",author:"Chris Sewell",author_email:"chrisj_sewell@hotmail.com",license:"MIT",home_page:"https://github.com/chrisjsewell/aiida-crystal17",classifiers:["Framework :: AiiDA","Programming Language :: Python","Programming Language :: Python :: 2.7","Programming Language :: Python :: 3.6","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics"],version:"0.11.0"},aiida_version:">=1.4.0,<2.0.0",entry_points:{"aiida.calculations":{"crystal17.basic":"aiida_crystal17.calculations.cry_basic:CryBasicCalculation","crystal17.doss":"aiida_crystal17.calculations.prop_doss:CryDossCalculation","crystal17.ech3":"aiida_crystal17.calculations.prop_ech3:CryEch3Calculation","crystal17.main":"aiida_crystal17.calculations.cry_main:CryMainCalculation","crystal17.newk":"aiida_crystal17.calculations.prop_newk:CryNewkCalculation","crystal17.ppan":"aiida_crystal17.calculations.prop_ppan:CryPpanCalculation"},"aiida.cmdline.data":{"crystal17.basis":"aiida_crystal17.cmndline.basis_set:basisset","crystal17.parse":"aiida_crystal17.cmndline.cmd_parser:parse","crystal17.symmetry":"aiida_crystal17.cmndline.symmetry:symmetry"},"aiida.data":{"crystal17.basisset":"aiida_crystal17.data.basis_set:BasisSetData","crystal17.gcube":"aiida_crystal17.data.gcube:GaussianCube","crystal17.kinds":"aiida_crystal17.data.kinds:KindData","crystal17.parameters":"aiida_crystal17.data.input_params:CryInputParamsData","crystal17.symmetry":"aiida_crystal17.data.symmetry:SymmetryData"},"aiida.groups":{"crystal17.basisset":"aiida_crystal17.data.basis_set:BasisSetFamily"},"aiida.parsers":{"crystal17.doss":"aiida_crystal17.parsers.cry_doss:CryDossParser","crystal17.ech3":"aiida_crystal17.parsers.cry_ech3:CryEch3Parser","crystal17.main":"aiida_crystal17.parsers.cry_main:CryMainParser","crystal17.newk":"aiida_crystal17.parsers.cry_newk:CryNewkParser","crystal17.ppan":"aiida_crystal17.parsers.cry_ppan:CryPpanParser"},"aiida.workflows":{"crystal17.main.base":"aiida_crystal17.workflows.crystal_main.base:CryMainBaseWorkChain","crystal17.properties":"aiida_crystal17.workflows.crystal_props.base:CryPropertiesWorkChain","crystal17.sym3d":"aiida_crystal17.workflows.symmetrise_3d_struct:Symmetrise3DStructure"},console_scripts:{mock_crystal17:"aiida_crystal17.tests.mock_crystal17:main",mock_properties17:"aiida_crystal17.tests.mock_properties17:main"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-crystal17
Collecting aiida-crystal17
+  Downloading aiida_crystal17-0.11.0-py3-none-any.whl.metadata (5.8 kB)
+INFO: pip is looking at multiple versions of aiida-crystal17 to determine which version is compatible with other requirements. This could take a while.
+  Downloading aiida-crystal17-0.4.0.tar.gz (120 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 120.8/120.8 kB 5.7 MB/s eta 0:00:00
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+ERROR: Cannot install aiida-crystal17==0.11.0 and aiida-crystal17==0.4.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-crystal17 0.11.0 depends on aiida-core<2.0.0 and >=1.4.0
+    aiida-crystal17 0.4.0 depends on aiida-core==0.12.2
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:6},{colorclass:"brown",text:"Parsers",count:5},{colorclass:"red",text:"Data",count:5},{colorclass:"green",text:"Workflows",count:3},{colorclass:"purple",text:"Console scripts",count:2},{colorclass:"orange",text:"Other (Data commands, Groups)",count:4}],pip_install_cmd:"pip install aiida-crystal17",is_installable:"False"},"aiida-cusp":{code_home:"https://github.com/aiida-cusp/aiida-cusp",documentation_url:"https://aiida-cusp.readthedocs.io",entry_point_prefix:"cusp",pip_url:"https://pypi.org/project/aiida-cusp",name:"aiida-cusp",package_name:"aiida_cusp",hosted_on:"github.com",metadata:{author:"Andreas Stamminger",author_email:"stammingera@gmail.com",version:"0.1.0b2",description:"Custodian based VASP Plugin for AiiDA",classifiers:["Development Status :: 4 - Beta","License :: OSI Approved :: MIT License","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Scientific/Engineering :: Physics","Topic :: Scientific/Engineering :: Chemistry","Environment :: Plugins","Framework :: AiiDA"]},aiida_version:">=1.3.0,<2.0.0",entry_points:{"aiida.data":{"cusp.kpoints":"aiida_cusp.data.inputs.vasp_kpoint:VaspKpointData","cusp.poscar":"aiida_cusp.data.inputs.vasp_poscar:VaspPoscarData","cusp.incar":"aiida_cusp.data.inputs.vasp_incar:VaspIncarData","cusp.potcar":"aiida_cusp.data.inputs.vasp_potcar:VaspPotcarData","cusp.vasprun":"aiida_cusp.data.outputs.vasp_vasprun:VaspVasprunData","cusp.outcar":"aiida_cusp.data.outputs.vasp_outcar:VaspOutcarData","cusp.contcar":"aiida_cusp.data.outputs.vasp_contcar:VaspContcarData","cusp.chgcar":"aiida_cusp.data.outputs.vasp_chgcar:VaspChgcarData","cusp.wavecar":"aiida_cusp.data.outputs.vasp_wavecar:VaspWavecarData","cusp.generic":"aiida_cusp.data.outputs.vasp_generic:VaspGenericData","cusp.potcarfile":"aiida_cusp.data.inputs.vasp_potcar:VaspPotcarFile"},"aiida.calculations":{"cusp.vasp":"aiida_cusp.calculators.vasp_calculation:VaspCalculation"},"aiida.parsers":{"cusp.default":"aiida_cusp.parsers.vasp_file_parser:VaspFileParser"},"aiida.cmdline.data":{potcar:"aiida_cusp.cli.potcar_cmd:potcar"}},commits_count:0,development_status:"beta",errors:[`E001: Failed to install plugin aiida-cusp
Collecting https://pypi.org/project/aiida-cusp
+  Downloading https://pypi.org/project/aiida-cusp (43 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 43.3/43.3 kB 6.0 MB/s eta 0:00:00
+  ERROR: Cannot unpack file /tmp/pip-unpack-0scpd51m/aiida-cusp (downloaded from /tmp/pip-req-build-vrjss5ft, content-type: text/html; charset=UTF-8); cannot detect archive format
+ERROR: Cannot determine archive format of /tmp/pip-req-build-vrjss5ft
+
`],warnings:["W010: Entry point 'potcar' does not start with prefix 'cusp.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:11},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install https://pypi.org/project/aiida-cusp",is_installable:"False"},"aiida-dataframe":{entry_point_prefix:"dataframe",plugin_info:"https://raw.github.com/janssenhenning/aiida-dataframe/main/pyproject.toml",code_home:"https://github.com/janssenhenning/aiida-dataframe",version_file:"https://raw.githubusercontent.com/janssenhenning/aiida-dataframe/main/aiida_dataframe/__init__.py",pip_url:"aiida-dataframe",documentation_url:"https://aiida-dataframe.readthedocs.io/en/latest/",name:"aiida-dataframe",package_name:"aiida_dataframe",hosted_on:"github.com",metadata:{release_date:"2023-05-05",description:"AiiDA data plugin for pandas DataFrame objects",author_email:"Henning Janßen ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"0.1.3"},aiida_version:">=1.0,<3",entry_points:{"aiida.cmdline.data":{dataframe:"aiida_dataframe.cli:data_cli"},"aiida.data":{"dataframe.frame":"aiida_dataframe.data.dataframe:PandasFrameData"}},commits_count:0,development_status:"beta",errors:[],warnings:[],summaryinfo:[{colorclass:"red",text:"Data",count:1},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install aiida-dataframe",is_installable:"True"},"aiida-ddec":{code_home:"https://github.com/lsmo-epfl/aiida-ddec",entry_point_prefix:"ddec",pip_url:"git+https://github.com/yakutovicha/aiida-ddec",name:"aiida-ddec",package_name:"aiida_ddec",hosted_on:"github.com",metadata:{description:"AiiDA plugin for DDEC code",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11"],author:"Aliaksandr Yakutovich",author_email:"aliaksandr.yakutovich@epfl.ch"},aiida_version:"~=2.3",entry_points:{"aiida.calculations":{ddec:{description:["AiiDA plugin for the ddec code that performs density derived"," electrostatic and chemical atomic population analysis."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters such as net charge, protocol, atomic densities path, ..."},{name:"charge_density_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote folder (for restarts and similar)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"structure_ddec",required:!0,valid_types:"CifData",info:"structure with DDEC charges"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The retrieved folder data node could not be accessed."},{status:101,message:"The retrieved folder does not contain an output file."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_ddec.calculations:DdecCalculation"}},"aiida.parsers":{ddec:"aiida_ddec.parsers:DdecParser"},"aiida.workflows":{"ddec.cp2k_ddec":{description:["A workchain that computes DDEC charges after a single-point"," DFT calculation using Cp2kBaseWorkChain"],spec:{inputs:[{name:"cp2k_base",required:!0,valid_types:"Data",info:""},{name:"ddec",required:!0,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"structure_ddec",required:!0,valid_types:"CifData",info:"structure with DDEC charges"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:903,message:"Error while parsing CP2K output"},{status:904,message:"Error while parsing DDEC output"}]},class:"aiida_ddec.workchains:Cp2kDdecWorkChain"}}},commits_count:0,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/yakutovicha/aiida-ddec",is_installable:"True"},"aiida-defects":{code_home:"https://github.com/epfl-theos/aiida-defects",entry_point_prefix:"defects",pip_url:"aiida-defects",plugin_info:"https://raw.githubusercontent.com/epfl-theos/aiida-defects/master/pyproject.toml",name:"aiida-defects",package_name:"aiida_defects",hosted_on:"github.com",metadata:{release_date:"2023-03-29",description:"AiiDA-Defects is a plugin for the AiiDA computational materials science framework, and provides tools and automated workflows for the study of defects in materials.",author:"The AiiDA-Defects developers",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"1.0.1"},aiida_version:">=2.0,<3",entry_points:{"aiida.data":{"defects.array.stability":"aiida_defects.data.data:StabilityData"},"aiida.workflows":{"defects.formation_energy.chemical_potential":"aiida_defects.formation_energy.chemical_potential.chemical_potential:ChemicalPotentialWorkchain","defects.formation_energy.corrections.gaussian_countercharge":"aiida_defects.formation_energy.corrections.gaussian_countercharge.gaussian_countercharge:GaussianCounterChargeWorkchain","defects.formation_energy.corrections.gaussian_countercharge.model_potential":"aiida_defects.formation_energy.corrections.gaussian_countercharge.model_potential.model_potential:ModelPotentialWorkchain","defects.formation_energy.corrections.point_countercharge":"aiida_defects.formation_energy.corrections.point_countercharge.point_countercharge:PointCounterChargeWorkchain","defects.formation_energy.potential_alignment":"aiida_defects.formation_energy.potential_alignment.potential_alignment:PotentialAlignmentWorkchain","defects.formation_energy.qe":"aiida_defects.formation_energy.formation_energy_qe:FormationEnergyWorkchainQE","defects.formation_energy.siesta":"aiida_defects.formation_energy.formation_energy_siesta:FormatonEnergyWorkchainSiesta"}},commits_count:0,development_status:"beta",errors:[`E003: Failed to fetch entry point metadata for package aiida_defects
Traceback (most recent call last):
+  File "/tmp/scripts/./bin/analyze_entrypoints.py", line 196, in 
+    cli()  # pylint: disable=no-value-for-parameter
+Failed to load entry point 'defects.formation_energy.corrections.point_countercharge':
+Traceback (most recent call last):
+  File "/opt/conda/lib/python3.10/site-packages/aiida/plugins/entry_point.py", line 278, in load_entry_point
+    loaded_entry_point = entry_point.load()
+  File "/opt/conda/lib/python3.10/site-packages/importlib_metadata/__init__.py", line 211, in load
+    module = import_module(match.group('module'))
+  File "/opt/conda/lib/python3.10/importlib/__init__.py", line 126, in import_module
+    return _bootstrap._gcd_import(name[level:], package, level)
+  File "", line 1050, in _gcd_import
+  File "", line 1027, in _find_and_load
+  File "", line 1006, in _find_and_load_unlocked
+  File "", line 688, in _load_unlocked
+  File "", line 883, in exec_module
+  File "", line 241, in _call_with_frames_removed
+  File "/opt/conda/lib/python3.10/site-packages/aiida_defects/formation_energy/corrections/point_countercharge/point_countercharge.py", line 19, in 
+    from aiida_defects.tools.structure_manipulation import get_spacegroup
+ModuleNotFoundError: No module named 'aiida_defects.tools'
+
+  File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1157, in __call__
+    return self.main(*args, **kwargs)
+  File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1078, in main
+    rv = self.invoke(ctx)
+  File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 1434, in invoke
+    return ctx.invoke(self.callback, **ctx.params)
+  File "/opt/conda/lib/python3.10/site-packages/click/core.py", line 783, in invoke
+    return __callback(*args, **kwargs)
+  File "/tmp/scripts/./bin/analyze_entrypoints.py", line 186, in cli
+    result[ep_group] = document_entry_point_group(ep_group)
+  File "/tmp/scripts/./bin/analyze_entrypoints.py", line 75, in document_entry_point_group
+    process_info = document_entry_point(entry_point_group, entry_point)
+  File "/tmp/scripts/./bin/analyze_entrypoints.py", line 94, in document_entry_point
+    plugin = load_entry_point(entry_point_group, entry_point)
+  File "/opt/conda/lib/python3.10/site-packages/aiida/plugins/entry_point.py", line 278, in load_entry_point
+    loaded_entry_point = entry_point.load()
+  File "/opt/conda/lib/python3.10/site-packages/importlib_metadata/__init__.py", line 213, in load
+    return functools.reduce(getattr, attrs, module)
+AttributeError: module 'aiida_defects.formation_energy.formation_energy_siesta' has no attribute 'FormatonEnergyWorkchainSiesta'. Did you mean: 'FormationEnergyWorkchainSiesta'?
+
`],warnings:[],summaryinfo:[{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:7}],pip_install_cmd:"pip install aiida-defects",is_installable:"True"},"aiida-diff":{code_home:"https://github.com/aiidateam/aiida-diff",development_status:"stable",documentation_url:"https://aiida-diff.readthedocs.io/",entry_point_prefix:"diff",pip_url:"git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0",name:"aiida-diff",package_name:"aiida_diff",hosted_on:"github.com",metadata:{description:"AiiDA demo plugin that wraps the `diff` executable for computing the difference between two files.",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Development Status :: 3 - Alpha","Framework :: AiiDA"],author:"The AiiDA Team"},aiida_version:">=2.5,<3",entry_points:{"aiida.data":{diff:"aiida_diff.data:DiffParameters"},"aiida.calculations":{diff:"aiida_diff.calculations:DiffCalculation"},"aiida.parsers":{diff:"aiida_diff.parsers:DiffParser"},"aiida.cmdline.data":{diff:"aiida_diff.cli:data_cli"}},commits_count:2,errors:[`E001: Failed to install plugin aiida-diff
Collecting aiida-diff-0.1.0a0
+  Cloning https://github.com/aiidateam/aiida-diff to /tmp/pip-install-ps0wxtvo/aiida-diff-0-1-0a0_cdad228f0c3e482f899eab94d021c5d2
+  Running command git clone --filter=blob:none --quiet https://github.com/aiidateam/aiida-diff /tmp/pip-install-ps0wxtvo/aiida-diff-0-1-0a0_cdad228f0c3e482f899eab94d021c5d2
+  Resolved https://github.com/aiidateam/aiida-diff to commit f9b151e11634b16e6a42d5445da23735415872e0
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+  WARNING: Generating metadata for package aiida-diff-0.1.0a0 produced metadata for project name aiida-diff. Fix your #egg=aiida-diff-0.1.0a0 fragments.
+Discarding git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0: Requested aiida-diff from git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0 has inconsistent name: expected 'aiida-diff-0-1-0a0', but metadata has 'aiida-diff'
+ERROR: Could not find a version that satisfies the requirement aiida-diff-0-1-0a0 (unavailable) (from versions: none)
+ERROR: No matching distribution found for aiida-diff-0-1-0a0 (unavailable)
+
`],warnings:["W005: Development status in classifiers (alpha) does not match development_status in metadata (stable)","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0",is_installable:"False"},"aiida-donothing":{code_home:"https://github.com/atztogo/aiida-donothing",entry_point_prefix:"donothing",name:"aiida-donothing",package_name:"aiida_donothing",hosted_on:"github.com",metadata:{author:"Atsushi Togo",author_email:"atz.togo@gmail.com",version:"0.1",description:"AiiDA calculation plugin for doing nothing",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=1.6.5,<2.0.0",entry_points:{"aiida.calculations":{"donothing.donothing":"aiida_donothing.calculations.donothing:DoNothingCalculation"},"aiida.parsers":{"donothing.donothing":"aiida_donothing.parsers.donothing:DoNothingParser"}},commits_count:0,development_status:"planning",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"See source code repository."},"aiida-dynamic-workflows":{code_home:"https://github.com/microsoft/aiida-dynamic-workflows",entry_point_prefix:"dynamic_workflows",name:"aiida-dynamic-workflows",package_name:"aiida_dynamic_workflows",hosted_on:"github.com",metadata:{},aiida_version:null,entry_points:{},commits_count:0,development_status:"planning",errors:[],warnings:["E001: Cannot fetch all data from PyPI and missing plugin_info key!","W002: AiiDA version not found"],summaryinfo:[],pip_install_cmd:"See source code repository."},"aiida-environ":{code_home:"https://github.com/environ-developers/aiida-environ",entry_point_prefix:"environ",pip_url:"git+https://github.com/environ-developers/aiida-environ",name:"aiida-environ",package_name:"aiida_environ",hosted_on:"github.com",metadata:{version:"1.0.0",description:"AiiDA plugin for Environ",classifiers:["Development Status :: 5 - Production/Stable","Programming Language :: Python","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],author_email:"materialab@unt.edu"},aiida_version:">=1.6",entry_points:{"aiida.calculations":{"environ.pw":{description:["`CalcJob` implementation for the pw.x code of Quantum ESPRESSO + Environ."],spec:{inputs:[{name:"environ_parameters",required:!0,valid_types:"Dict",info:"The input parameters that are to be used to construct the input file."},{name:"kpoints",required:!0,valid_types:"KpointsData",info:"kpoint mesh or kpoint path"},{name:"parameters",required:!0,valid_types:"Dict",info:"The input parameters that are to be used to construct the input file."},{name:"pseudos",required:!0,valid_types:"UpfData, UpfData",info:"A mapping of `UpfData` nodes onto the kind name to which they should apply."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"external_charges",required:!1,valid_types:"EnvironChargeData, NoneType",info:"External charges"},{name:"hubbard_file",required:!1,valid_types:"SinglefileData, NoneType",info:"SinglefileData node containing the output Hubbard parameters from a HpCalculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parallelization",required:!1,valid_types:"Dict, NoneType",info:`Parallelization options. The following flags are allowed: +npool : The number of 'pools', each taking care of a group of k-points. +nband : The number of 'band groups', each taking care of a group of Kohn-Sham orbitals. +ntg : The number of 'task groups' across which the FFT planes are distributed. +ndiag : The number of 'linear algebra groups' used when parallelizing the subspace diagonalization / iterative orthonormalization. By default, no parameter is passed to Quantum ESPRESSO, meaning it will use its default.`},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"An optional working directory of a previously completed calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job and the parsing are performed."},{name:"vdw_table",required:!1,valid_types:"SinglefileData, NoneType",info:"Optional van der Waals table contained in a `SinglefileData`."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The `output_parameters` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_atomic_occupations",required:!1,valid_types:"Dict",info:""},{name:"output_band",required:!1,valid_types:"BandsData",info:"The `output_band` output node of the successful calculation if present."},{name:"output_kpoints",required:!1,valid_types:"KpointsData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The `output_structure` output node of the successful calculation if present."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The retrieved folder did not contain the required XML file."},{status:304,message:"The retrieved folder contained multiple XML files."},{status:305,message:"Both the stdout and XML output files could not be read or parsed."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:320,message:"The XML output file could not be read."},{status:321,message:"The XML output file could not be parsed."},{status:322,message:"The XML output file has an unsupported format."},{status:340,message:"The calculation stopped prematurely because it ran out of walltime but the job was killed by the scheduler before the files were safely written to disk for a potential restart."},{status:350,message:"The parser raised an unexpected exception: {exception}"},{status:360,message:"The code failed in finding a valid reciprocal lattice vector."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:410,message:"The electronic minimization cycle did not reach self-consistency."},{status:461,message:"The code failed with negative dexx in the exchange calculation."},{status:462,message:"The code failed during the cholesky factorization."},{status:463,message:"Too many bands failed to converge during the diagonalization."},{status:464,message:"The S matrix was found to be not positive definite."},{status:465,message:"The `zhegvd` failed in the PPCG diagonalization."},{status:466,message:"The `[Q, R] = qr(X, 0)` failed in the PPCG diagonalization."},{status:467,message:"The eigenvector failed to converge."},{status:468,message:"The factorization in the Broyden routine failed."},{status:481,message:'The k-point parallelization "npools" is too high, some nodes have no k-points.'},{status:500,message:"The ionic minimization cycle did not converge for the given thresholds."},{status:501,message:"Then ionic minimization cycle converged but the thresholds are exceeded in the final SCF."},{status:502,message:"The ionic minimization cycle did not converge after the maximum number of steps."},{status:503,message:"The ionic minimization cycle did not finish because the calculation was interrupted but a partial trajectory and output structure was successfully parsed which can be used for a restart."},{status:510,message:"The electronic minimization cycle failed during an ionic minimization cycle."},{status:511,message:"The ionic minimization cycle converged, but electronic convergence was not reached in the final SCF."},{status:520,message:"The ionic minimization cycle terminated prematurely because of two consecutive failures in the BFGS algorithm."},{status:521,message:"The ionic minimization cycle terminated prematurely because of two consecutive failures in the BFGS algorithm and electronic convergence failed in the final SCF."},{status:531,message:"The electronic minimization cycle did not reach self-consistency."},{status:541,message:"The variable cell optimization broke the symmetry of the k-points."},{status:542,message:"The cell relaxation caused a significant volume contraction and there is not enough space allocated for radial FFT."},{status:710,message:"The electronic minimization cycle did not reach self-consistency, but `scf_must_converge` is `False` and/or `electron_maxstep` is 0."}]},class:"aiida_environ.calculations.pw:EnvPwCalculation"},"environ.finite":{description:["Returns finite differences for a PK list and test settings.",""," Inputs:"," pk_list: aiida.orm.List"," test_settings: aiida.orm.Dict"," environ: aiida.orm.Bool",""," Outputs:"," data: aiida.orm.Dict"],spec:{inputs:[{name:"pk_list",required:!0,valid_types:"List",info:""},{name:"test_settings",required:!0,valid_types:"Dict",info:""},{name:"environ",required:!1,valid_types:"Bool",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_environ.calculations.finite:calculate_finite_differences"}},"aiida.data":{"environ.charges":"aiida_environ.data.charge:EnvironChargeData"},"aiida.parsers":{"environ.pw":"aiida_environ.parsers.pw:EnvPwParser"},"aiida.workflows":{"environ.pw.base":"aiida_environ.workflows.pw.base:EnvPwBaseWorkChain","environ.pw.relax":"aiida_environ.workflows.pw.relax:EnvPwRelaxWorkChain","environ.pw.solvation":"aiida_environ.workflows.pw.solvation:PwSolvationWorkChain","environ.pw.force_test":"aiida_environ.workflows.pw.force_test:EnvPwForceTestWorkChain"}},commits_count:0,development_status:"stable",errors:[],warnings:["W003: Missing classifier 'Framework :: AiiDA'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:4}],pip_install_cmd:"pip install git+https://github.com/environ-developers/aiida-environ",is_installable:"True"},"aiida-eon":{code_home:"https://github.com/HaoZeke/aiida-eon",entry_point_prefix:"eon",name:"aiida-eon",package_name:"aiida_eon",hosted_on:"github.com",metadata:{},aiida_version:null,entry_points:{},commits_count:0,development_status:"planning",errors:[],warnings:["E001: Cannot fetch all data from PyPI and missing plugin_info key!","W002: AiiDA version not found"],summaryinfo:[],pip_install_cmd:"See source code repository."},"aiida-eonclient":{code_home:"https://github.com/HaoZeke/aiida-eonclient",entry_point_prefix:"eonclient",name:"aiida-eonclient",package_name:"aiida_eonclient",hosted_on:"github.com",metadata:{},aiida_version:null,entry_points:{},commits_count:0,development_status:"planning",errors:[],warnings:["E001: Cannot fetch all data from PyPI and missing plugin_info key!","W002: AiiDA version not found"],summaryinfo:[],pip_install_cmd:"See source code repository."},"aiida-fenics":{code_home:"https://github.com/broeder-j/aiida-fenics/tree/master",entry_point_prefix:"fenics",pip_url:"git+https://github.com/broeder-j/aiida-fenics",name:"aiida-fenics",package_name:"aiida_fenics",hosted_on:"github.com",metadata:{author:"Sebastiaan P. Huber",author_email:"mail@sphuber.net",version:"0.1.0",description:"AiiDA plugin for the FEniCS computing platform for solving partial differential equations.",classifiers:["Development Status :: 2 - Pre-Alpha","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: POSIX :: Linux","Operating System :: MacOS :: MacOS X","Programming Language :: Python","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"]},aiida_version:"~=1.6",entry_points:{"aiida.workflows":{"fenics.geometry_poisson":"aiida_fenics.workflows.geometry_poisson:GeometryPoisson"}},commits_count:0,development_status:"pre-alpha",errors:[`E001: Failed to install plugin aiida-fenics
Collecting git+https://github.com/broeder-j/aiida-fenics
+  Cloning https://github.com/broeder-j/aiida-fenics to /tmp/pip-req-build-u53c3kzm
+  Running command git clone --filter=blob:none --quiet https://github.com/broeder-j/aiida-fenics /tmp/pip-req-build-u53c3kzm
+  Resolved https://github.com/broeder-j/aiida-fenics to commit eae053f7c82889a15fe19c3909d303f923fcf79e
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-fenics to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-fenics==0.1.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-fenics 0.1.0 depends on aiida-core~=1.6
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:[],summaryinfo:[{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/broeder-j/aiida-fenics",is_installable:"False"},"aiida-fhiaims":{entry_point_prefix:"fhiaims",code_home:"https://gitlab.com/FHI-aims-club/utilities/aiida-fhiaims",pip_url:"git+https://gitlab.com/FHI-aims-club/utilities/aiida-fhiaims",version_file:"https://gitlab.com/FHI-aims-club/utilities/aiida-fhiaims/-/raw/master/aiida_fhiaims/__init__.py",name:"aiida-fhiaims",package_name:"aiida_fhiaims",hosted_on:"gitlab.com",metadata:{description:"An AiiDA plugin for FHI-aims code",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Development Status :: 3 - Alpha","Framework :: AiiDA"],author:"Andrey Sobolev",author_email:"andrey.n.sobolev@gmail.com"},aiida_version:">=2.0,<3",entry_points:{"aiida.cmdline.data":{"fhiaims.species_defaults":"aiida_fhiaims.cli.species_defaults:species_defaults"},"aiida.data":{"fhiaims.species_file":"aiida_fhiaims.data.species_file:BasisFile","fhiaims.parameters":"aiida_fhiaims.data.parameters:AimsParameters"},"aiida.groups":{"fhiaims.species_family":"aiida_fhiaims.data.species_family:BasisFamily"},"aiida.calculations":{fhiaims:{description:["AiiDA calculation plugin wrapping the aims.x executable."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"AimsParameters",info:"FHI-aims parameters dictionary"},{name:"structure",required:!0,valid_types:"StructureData",info:"Atomic structure to be calculated"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"fhiaims",required:!0,valid_types:"",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_fhiaims.calculations:AimsCalculation"}},"aiida.parsers":{"fhiaims.json":"aiida_fhiaims.parsers.aims_json:AimsJSONParser","fhiaims.out":"aiida_fhiaims.parsers.aims_out:AimsOutParser"}},commits_count:2,development_status:"alpha",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:2},{colorclass:"orange",text:"Other (Data commands, Groups)",count:2}],pip_install_cmd:"pip install git+https://gitlab.com/FHI-aims-club/utilities/aiida-fhiaims",is_installable:"True"},"aiida-firecrest":{code_home:"https://github.com/aiidateam/aiida-firecrest",entry_point_prefix:"firecrest",pip_url:"aiida-firecrest",plugin_info:"https://raw.githubusercontent.com/aiidateam/aiida-firecrest/main/pyproject.toml",name:"aiida-firecrest",package_name:"aiida_firecrest",hosted_on:"github.com",metadata:{release_date:"2022-01-14",description:"AiiDA Transport/Scheduler plugins for interfacing with FirecREST.",author_email:"Chris Sewell ",classifiers:["Development Status :: 3 - Alpha","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Programming Language :: Python :: Implementation :: CPython","Topic :: Software Development :: Libraries :: Python Modules"],version:"0.1.0a1"},aiida_version:"<2",entry_points:{"aiida.schedulers":{firecrest:"aiida_firecrest.scheduler:FirecrestScheduler"},"aiida.transports":{firecrest:"aiida_firecrest.transport:FirecrestTransport"},console_scripts:{"aiida-firecrest-cli":"aiida_firecrest.cli:main"}},commits_count:10,development_status:"alpha",errors:[`E001: Failed to install plugin aiida-firecrest
Collecting aiida-firecrest
+  Downloading aiida_firecrest-0.1.0a1-py3-none-any.whl.metadata (4.2 kB)
+INFO: pip is looking at multiple versions of aiida-firecrest to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-firecrest==0.1.0a1 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-firecrest 0.1.0a1 depends on aiida-core<2
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:[],summaryinfo:[{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Schedulers, Transports)",count:2}],pip_install_cmd:"pip install --pre aiida-firecrest",is_installable:"False"},"aiida-fireworks-scheduler":{code_home:"https://github.com/zhubonan/aiida-fireworks-scheduler",development_status:"beta",documentation_url:"https://aiida-fireworks-scheduler.readthedocs.io",entry_point_prefix:"fireworks_scheduler",pip_url:"git+https://github.com/zhubonan/aiida-fireworks-scheduler",name:"aiida-fireworks-scheduler",package_name:"aiida_fireworks_scheduler",hosted_on:"github.com",metadata:{author:"Bonan Zhu",author_email:"zhubonan@outlook.com",version:"1.2.0",description:"AiiDA plugin to allow using `fireworks` as the execution engine for `CalcJob`.",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:null,entry_points:{"aiida.schedulers":{fireworks:"aiida_fireworks_scheduler.fwscheduler:FwScheduler","fireworks_scheduler.default":"aiida_fireworks_scheduler.fwscheduler:FwScheduler","fireworks_scheduler.keepenv":"aiida_fireworks_scheduler.fwscheduler:FwSchedulerKeepEnv"},"aiida.cmdline.data":{"fireworks-scheduler":"aiida_fireworks_scheduler.cmdline:fw_cli"},console_scripts:{arlaunch:"aiida_fireworks_scheduler.scripts.arlaunch_run:arlaunch"}},commits_count:0,errors:[],warnings:["W002: AiiDA version not found","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'fireworks' does not start with prefix 'fireworks_scheduler.'","W010: Entry point 'fireworks-scheduler' does not start with prefix 'fireworks_scheduler.'"],summaryinfo:[{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Data commands, Schedulers)",count:4}],pip_install_cmd:"pip install git+https://github.com/zhubonan/aiida-fireworks-scheduler",is_installable:"True"},"aiida-fleur":{code_home:"https://github.com/JuDFTteam/aiida-fleur/tree/develop",development_status:"stable",documentation_url:"https://aiida-fleur.readthedocs.io/",entry_point_prefix:"fleur",pip_url:"aiida-fleur",plugin_info:"https://raw.github.com/JuDFTteam/aiida-fleur/develop/setup.json",name:"aiida-fleur",package_name:"aiida_fleur",hosted_on:"github.com",metadata:{release_date:"2023-05-03",description:"AiiDA Plugin for running the FLEUR code and its input generator. Also includes high-level workchains and utilities",author_email:"The JuDFT team ",classifiers:["Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Physics"],version:"2.0.0"},aiida_version:">=2.0.1,<3.0.0",entry_points:{"aiida.calculations":{"fleur.fleur":{description:["A CalcJob class that represents FLEUR DFT calculation."," For more information about the FLEUR-code family go to http://www.flapw.de/"],spec:{inputs:[{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:"Use a FleurinpData node that specifies the input parametersusually copy from the parent calculation, basically makesthe inp.xml file visible in the db and makes sure it has the files needed."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote or local repository folder as parent folder (also for restarts and similar). It should contain all the needed files for a Fleur calc, only edited files should be uploaded from the repository."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"This parameter data node is used to specify for some advanced features how the plugin behaves. You can add filesthe retrieve list, or add command line switches, for all available features here check the documentation."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"error_params",required:!1,valid_types:"Dict",info:""},{name:"output_parameters",required:!1,valid_types:"Dict",info:""},{name:"output_params_complex",required:!1,valid_types:"Dict",info:""},{name:"relax_parameters",required:!1,valid_types:"Dict",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"No retrieved folder found."},{status:301,message:"One of the output files can not be opened."},{status:302,message:"FLEUR calculation failed for unknown reason."},{status:303,message:"XML output file was not found."},{status:304,message:"Parsing of XML output file failed."},{status:305,message:"Parsing of relax XML output file failed."},{status:310,message:"FLEUR calculation failed due to lack of memory."},{status:311,message:"FLEUR calculation failed because an atom spilled to thevacuum during relaxation"},{status:312,message:"FLEUR calculation failed due to MT overlap."},{status:313,message:"Overlapping MT-spheres during relaxation."},{status:314,message:"Problem with cdn is suspected. Consider removing cdn"},{status:315,message:"The LDA+U density matrix contains invalid elements."},{status:316,message:"Calculation failed due to time limits."},{status:318,message:"Calculation failed due to missing dependency ({name}) for given calculation."}]},class:"aiida_fleur.calculation.fleur:FleurCalculation"},"fleur.inpgen":{description:["JobCalculationClass for the inpgen, which is a preprocessor for a FLEUR calculation."," For more information about produced files and the FLEUR-code family, go to http://www.flapw.de/."],spec:{inputs:[{name:"structure",required:!0,valid_types:"StructureData",info:"Choose the input structure to use"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Use a node that specifies the input parameters for the namelists"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"This parameter data node is used to specify for some advanced features how the plugin behaves. You can add filesthe retrieve list, or add command line switches, for all available features here check the documentation."}],outputs:[{name:"fleurinp",required:!0,valid_types:"FleurinpData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"No retrieved folder found."},{status:301,message:"One of the output files can not be opened."},{status:306,message:"XML input file was not found."},{status:307,message:"Some required files were not retrieved."},{status:308,message:"During parsing: FleurinpData could not be initialized, see log. "},{status:309,message:"During parsing: FleurinpData failed validation."},{status:310,message:"The profile {profile} is not known to the used inpgen code"}]},class:"aiida_fleur.calculation.fleurinputgen:FleurinputgenCalculation"}},"aiida.data":{"fleur.fleurinp":"aiida_fleur.data.fleurinp:FleurinpData"},"aiida.parsers":{"fleur.fleurinpgenparser":"aiida_fleur.parsers.fleur_inputgen:Fleur_inputgenParser","fleur.fleurparser":"aiida_fleur.parsers.fleur:FleurParser"},"aiida.workflows":{"fleur.banddos":{description:["This workflow calculated a bandstructure from a Fleur calculation",""," :Params: a Fleurcalculation node"," :returns: Success, last result node, list with convergence behavior"],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"scf",required:!1,valid_types:"Data",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"banddos_calc",required:!0,valid_types:"",info:""},{name:"output_banddos_wc_para",required:!0,valid_types:"Dict",info:""},{name:"output_banddos_wc_bands",required:!1,valid_types:"BandsData",info:""},{name:"output_banddos_wc_dos",required:!1,valid_types:"XyData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:233,message:"Invalid code node specified, check inpgen and fleur code nodes."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:334,message:"SCF calculation failed."},{status:335,message:"Found no SCF calculation remote repository."}]},class:"aiida_fleur.workflows.banddos:FleurBandDosWorkChain"},"fleur.base":{description:["Workchain to run a FLEUR calculation with automated error handling and restarts"],spec:{inputs:[{name:"options",required:!0,valid_types:"Dict",info:"Optional parameters to set up computational details."},{name:"add_comp_para",required:!1,valid_types:"Dict",info:"Gives additional control over computational parametersonly_even_MPI: set to true if you want to suppress odd number of MPI processes in parallelisation.This might speedup a calculation for machines having even number of sockets per node.max_queue_nodes: maximal number of nodes allowed on the remote machine. Used only to automatically solve some FLEUR failures.max_queue_wallclock_sec: maximal wallclock time allowed on the remote machine. Used only to automatically solve some FLEUR failures."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"description",required:!1,valid_types:"str, NoneType",info:"Calculation description."},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:"Use a FleurinpData node that specifies the input parametersusually copy from the parent calculation, basically makesthe inp.xml file visible in the db and makes sure it has the files needed."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"label",required:!1,valid_types:"str, NoneType",info:"Calculation label."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a remote or local repository folder as parent folder (also for restarts and similar). It should contain all the needed files for a Fleur calc, only edited files should be uploaded from the repository."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"This parameter data node is used to specify for some advanced features how the plugin behaves. You can add filesthe retrieve list, or add command line switches, for all available features here check the documentation."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"error_params",required:!1,valid_types:"Dict",info:""},{name:"output_parameters",required:!1,valid_types:"Dict",info:""},{name:"output_params_complex",required:!1,valid_types:"Dict",info:""},{name:"relax_parameters",required:!1,valid_types:"Dict",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:311,message:"FLEUR calculation failed because an atom spilled to thevacuum during relaxation"},{status:313,message:"Overlapping MT-spheres during relaxation."},{status:388,message:"Computational resources are not optimal."},{status:389,message:"Computational resources are not optimal."},{status:390,message:"Computational resources are not optimal."},{status:399,message:"FleurCalculation failed and FleurBaseWorkChain has no strategy to resolve this"},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_fleur.workflows.base_fleur:FleurBaseWorkChain"},"fleur.base_relax":{description:["Workchain to run Relax WorkChain with automated error handling and restarts"],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"description",required:!1,valid_types:"str, NoneType",info:"Calculation description."},{name:"final_scf",required:!1,valid_types:"Data",info:""},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"label",required:!1,valid_types:"str, NoneType",info:"Calculation label."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"last_scf",required:!0,valid_types:"",info:""},{name:"optimized_structure",required:!0,valid_types:"StructureData",info:""},{name:"output_relax_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:399,message:"FleurRelaxWorkChain failed and FleurBaseRelaxWorkChain has no strategy to resolve this"},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_fleur.workflows.base_relax:FleurBaseRelaxWorkChain"},"fleur.cfcoeff":{description:["Workflow for calculating rare-earth crystal field coefficients"],spec:{inputs:[{name:"metadata",required:!1,valid_types:"",info:""},{name:"orbcontrol",required:!1,valid_types:"Data",info:""},{name:"scf",required:!1,valid_types:"Data",info:""},{name:"scf_rare_earth_analogue",required:!1,valid_types:"Data",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_cfcoeff_wc_para",required:!0,valid_types:"Dict",info:""},{name:"output_cfcoeff_wc_charge_densities",required:!1,valid_types:"XyData",info:""},{name:"output_cfcoeff_wc_potentials",required:!1,valid_types:"XyData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:345,message:"Convergence scf workflow failed."},{status:451,message:"Convergence orbcontrol workflow failed."},{status:452,message:"CF calculation failed."}]},class:"aiida_fleur.workflows.cfcoeff:FleurCFCoeffWorkChain"},"fleur.corehole":{description:["Turn key solution for a corehole calculation with the FLEUR code."," Has different protocols for different core-hole types (valence, charge).",""," Calculates supercells. Extracts binding energies"," for certain corelevels from the total energy differences a the calculation with"," corehole and without.",""," Documentation:"," See help for details.",""," Two paths are possible:",""," (1) Start from a structure -> workchains run inpgen first (recommended)"," (2) Start from a Fleurinp data object",""," Also it is recommended to provide a calc parameter node for the structure",""," :param wf_parameters: Dict node, specify, resources and what should be calculated"," :param structure: structureData node, crystal structure"," :param calc_parameters: Dict node, inpgen parameters for the crystal structure"," :param fleurinp: fleurinpData node,"," :param inpgen: Code node,"," :param fleur: Code node,",""," :return: output_corehole_wc_para Dict node, successful=True if no error",""," :uses workchains: fleur_scf_wc, fleur_relax_wc"," :uses calcfunctions: supercell, create_corehole_result_node, prepare_struc_corehole_wf"],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"inpgen",required:!0,valid_types:"Code",info:""},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_corehole_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:1,message:"The input resources are invalid."},{status:2,message:"The process failed with legacy failure mode."},{status:2,message:"Input resources are missing."},{status:3,message:"The code provided is invalid, or not of the right kind."},{status:4,message:"Inpgen calculation FAILED, check output"},{status:5,message:"Changing of the FLEURINP data went wrong, check log."},{status:6,message:"The FLEUR input file for the calculation did not validate."},{status:7,message:"At least one FLEUR calculation FAILED, check the output and log."},{status:8,message:"At least one FLEUR calculation did not/could not reach thedesired convergece Criteria, with the current parameters."},{status:9,message:"Something went wrong in the determiation what coreholes to calculate, probably the input format was not correct. Check log."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_fleur.workflows.corehole:FleurCoreholeWorkChain"},"fleur.create_magnetic":{description:["This workflow creates relaxed magnetic film on a substrate."],spec:{inputs:[{name:"distance_suggestion",required:!1,valid_types:"Dict, NoneType",info:""},{name:"eos",required:!1,valid_types:"Data",info:""},{name:"eos_output",required:!1,valid_types:"Dict, NoneType",info:""},{name:"interlayer_dist",required:!1,valid_types:"Dict, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"optimized_structure",required:!1,valid_types:"StructureData, NoneType",info:""},{name:"relax",required:!1,valid_types:"Data",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"magnetic_structure",required:!0,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:380,message:"Specified substrate has to be bcc or fcc."},{status:382,message:"Relaxation calculation failed."},{status:383,message:"EOS WorkChain failed."}]},class:"aiida_fleur.workflows.create_magnetic_film:FleurCreateMagneticWorkChain"},"fleur.dmi":{description:["This workflow calculates DMI energy dispersion of a structure."],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"scf",required:!1,valid_types:"Data",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_dmi_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:233,message:"Invalid code node specified, check inpgen and fleur code nodes."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:334,message:"Reference calculation failed."},{status:335,message:"Found no reference calculation remote repository."},{status:336,message:"Force theorem calculation failed."}]},class:"aiida_fleur.workflows.dmi:FleurDMIWorkChain"},"fleur.dos":{description:["DEPRECATED: Use FleurBandDosWorkChain instead (entrypoint fleur.banddos)"," This workflow calculated a DOS from a Fleur calculation",""," :Params: a Fleurcalculation node"," :returns: Success, last result node, list with convergence behavior",""," wf_parameters: { 'tria', 'nkpts', 'sigma', 'emin', 'emax'}"," defaults : tria = True, nkpts = 800, sigma=0.005, emin= -0.3, emax = 0.8"],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote_data",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_fleur.workflows.dos:fleur_dos_wc"},"fleur.eos":{description:["This workflow calculates the equation of states of a structure."," Calculates several unit cells with different volumes."," A Birch_Murnaghan equation of states fit determines the Bulk modulus and the"," groundstate volume of the cell.",""," :params wf_parameters: Dict node, optional 'wf_parameters', protocol specifying parameter dict"," :params structure: StructureData node, 'structure' crystal structure"," :params calc_parameters: Dict node, optional 'calc_parameters' parameters for inpgen"," :params inpgen: Code node,"," :params fleur: Code node,","",""," :return output_eos_wc_para: Dict node, contains relevant output information."," about general succeed, fit results and so on."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_eos_wc_para",required:!0,valid_types:"Dict",info:""},{name:"output_eos_wc_structure",required:!0,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:400,message:"At least one of the SCF sub processes did not finish successfully."}]},class:"aiida_fleur.workflows.eos:FleurEosWorkChain"},"fleur.init_cls":{description:["Turn key solution for the calculation of core level shift"],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"inpgen",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_initial_cls_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_fleur.workflows.initial_cls:FleurInitialCLSWorkChain"},"fleur.mae":{description:["This workflow calculates the Magnetic Anisotropy Energy of a structure."],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"scf",required:!1,valid_types:"Data",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_mae_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:233,message:"Invalid code node specified, check inpgen and fleur code nodes."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:334,message:"Reference calculation failed."},{status:335,message:"Found no reference calculation remote repository."},{status:336,message:"Force theorem calculation failed."}]},class:"aiida_fleur.workflows.mae:FleurMaeWorkChain"},"fleur.mae_conv":{description:["This workflow calculates the Magnetic Anisotropy Energy of a structure."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_mae_conv_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:343,message:"Convergence MAE calculation failed for all SQAs."},{status:344,message:"Convergence MAE calculation failed for some SQAs."}]},class:"aiida_fleur.workflows.mae_conv:FleurMaeConvWorkChain"},"fleur.orbcontrol":{description:["Workchain for determining the groundstate density matrix in an DFT+U"," calculation. This is done in 2 or 3 steps:",""," 1. Converge the system without DFT+U (a converged calculation can be"," provided to skip this step)"," 2. A fixed number of iterations is run with fixed density matrices"," either generated as all distinct permutations for the given occupations"," or the explicitly given configurations"," 3. The system and density matrix is relaxed",""," :param wf_parameters: (Dict), Workchain Specifications"," :param scf_no_ldau: (Dict), Inputs to a FleurScfWorkChain providing the initial system"," either converged or staring from a structure"," :param scf_with_ldau: (Dict), Inputs to a FleurScfWorkChain. Only the wf_parameters are valid"," :param fleurinp: (FleurinpData) FleurinpData to start from if no SCF should be done"," :param remote: (RemoteData) RemoteData to start from if no SCF should be done"," :param structure: (StructureData) Structure to start from if no SCF should be done"," :param calc_parameters: (Dict), Inpgen Parameters"," :param settings: (Dict), additional settings for e.g retrieving files"," :param options: (Dict), Options for the submission of the jobs"," :param inpgen: (Code)"," :param fleur: (Code)"],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"fixed_remotes",required:!1,valid_types:"RemoteData",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"inpgen",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"options_inpgen",required:!1,valid_types:"Dict, NoneType",info:""},{name:"relaxed_remotes",required:!1,valid_types:"RemoteData",info:""},{name:"remote",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"scf_no_ldau",required:!1,valid_types:"Data",info:"Inputs for SCF Workchain before adding LDA+U"},{name:"scf_with_ldau",required:!1,valid_types:"Data",info:"Inputs for SCF Workchain after the LDA+U matrix was fixed"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"settings_inpgen",required:!1,valid_types:"Dict, NoneType",info:""},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"groundstate_scf",required:!0,valid_types:"",info:""},{name:"output_orbcontrol_wc_para",required:!0,valid_types:"Dict",info:""},{name:"groundstate_denmat",required:!1,valid_types:"SinglefileData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:233,message:"Input codes do not correspond to fleur or inpgen respectively."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:342,message:"Convergence LDA+U calculation failed for some Initial configurations."},{status:343,message:"Convergence LDA+U calculation failed for all Initial configurations."},{status:360,message:"Inpgen calculation failed."},{status:450,message:"Convergence workflow without LDA+U failed."}]},class:"aiida_fleur.workflows.orbcontrol:FleurOrbControlWorkChain"},"fleur.relax":{description:["This workflow performs structure optimization."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:""},{name:"final_scf",required:!1,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"last_scf",required:!0,valid_types:"",info:""},{name:"optimized_structure",required:!0,valid_types:"StructureData",info:""},{name:"output_relax_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"If you want to run a final scf inpgen has to be there."},{status:311,message:"FLEUR calculation failed because an atom spilled to thevacuum during relaxation"},{status:313,message:"Overlapping MT-spheres during relaxation."},{status:350,message:"Optimization cycle did not lead to convergence of forces."},{status:351,message:"SCF Workchains failed for some reason."},{status:352,message:"Found no relaxed structure info in the output of SCF"},{status:353,message:"Found no SCF output"},{status:354,message:"Force is small, switch to BFGS"}]},class:"aiida_fleur.workflows.relax:FleurRelaxWorkChain"},"fleur.relax_torque":{description:["This workflow performs spin structure optimization."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:""},{name:"final_scf",required:!1,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_relax_torque_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"If you want to run a final scf inpgen has to be there."},{status:350,message:"Optimization cycle did not lead to convergence."},{status:351,message:"An SCF Workchain failed for some reason."}]},class:"aiida_fleur.workflows.relax_torque:FleurRelaxTorqueWorkChain"},"fleur.scf":{description:["Workchain for converging a FLEUR calculation (SCF).",""," It converges the charge density, total energy or the largest force."," Two paths are possible:",""," (1) Start from a structure and run the inpgen first optional with calc_parameters"," (2) Start from a Fleur calculation, with optional remoteData",""," :param wf_parameters: (Dict), Workchain Specifications"," :param structure: (StructureData), Crystal structure"," :param calc_parameters: (Dict), Inpgen Parameters"," :param fleurinp: (FleurinpData), to start with a Fleur calculation"," :param remote_data: (RemoteData), from a Fleur calculation"," :param inpgen: (Code)"," :param fleur: (Code)",""," :return: output_scf_wc_para (Dict), Information of workflow results"," like Success, last result node, list with convergence behavior"],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"inpgen",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote_data",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"settings_inpgen",required:!1,valid_types:"Dict, NoneType",info:""},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"fleurinp",required:!0,valid_types:"FleurinpData",info:""},{name:"last_calc",required:!0,valid_types:"",info:""},{name:"output_scf_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:233,message:"Input codes do not correspond to fleur or inpgen respectively."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:360,message:"Inpgen calculation failed."},{status:361,message:"Fleur calculation failed."},{status:362,message:"SCF cycle did not lead to convergence."}]},class:"aiida_fleur.workflows.scf:FleurScfWorkChain"},"fleur.ssdisp":{description:["This workflow calculates spin spiral dispersion of a structure."],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"fleurinp",required:!1,valid_types:"FleurinpData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"scf",required:!1,valid_types:"Data",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_ssdisp_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:231,message:"Invalid input configuration."},{status:233,message:"Invalid code node specified, check inpgen and fleur code nodes."},{status:235,message:"Input file modification failed."},{status:236,message:"Input file was corrupted after user's modifications."},{status:334,message:"Reference calculation failed."},{status:335,message:"Found no reference calculation remote repository."},{status:336,message:"Force theorem calculation failed."}]},class:"aiida_fleur.workflows.ssdisp:FleurSSDispWorkChain"},"fleur.ssdisp_conv":{description:["This workflow calculates the Spin Spiral Dispersion of a structure."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_ssdisp_conv_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:230,message:"Invalid workchain parameters."},{status:340,message:"Convergence SSDisp calculation failed for all q-vectors."},{status:341,message:"Convergence SSDisp calculation failed for some q-vectors."}]},class:"aiida_fleur.workflows.ssdisp_conv:FleurSSDispConvWorkChain"},"fleur.strain":{description:["This workflow calculates the deformation potential a structure = -BdEg/dP = d(Eg)/d(ln(V))."," Calculates several unit cells with different volumes."," A Birch_Murnaghan equation of states fit determines the Bulk modulus(B) and the"," ground-state volume of the cell.",""," :params wf_parameters: Dict node, optional 'wf_parameters', protocol specifying parameter dict"," :params structure: StructureData node, 'structure' crystal structure"," :params calc_parameters: Dict node, optional 'calc_parameters' parameters for inpgen"," :params inpgen: Code node,"," :params fleur: Code node,","",""," :return output_strain_wc_para: Dict node, contains relevant output information."," about general succeed, fit results and so on."],spec:{inputs:[{name:"fleur",required:!0,valid_types:"Code",info:""},{name:"inpgen",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_strain_wc_para",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:331,message:"Invalid code node specified, check inpgen and fleur code nodes."}]},class:"aiida_fleur.workflows.strain:FleurStrainWorkChain"}},console_scripts:{"aiida-fleur":"aiida_fleur.cmdline:cmd_root"}},commits_count:18,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:19},{colorclass:"purple",text:"Console scripts",count:1}],pip_install_cmd:"pip install aiida-fleur",is_installable:"True"},"aiida-flexpart":{code_home:"https://github.com/aiidaplugins/aiida-flexpart",entry_point_prefix:"flexpart",pip_url:"git+https://github.com/aiidaplugins/aiida-flexpart",name:"aiida-flexpart",package_name:"aiida_flexpart",hosted_on:"github.com",metadata:{version:"0.2.1",description:"AiiDA plugin for the FLEXPART code (simulation of atmospheric transport processes).",classifiers:["Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA","Development Status :: 3 - Alpha","Topic :: Scientific/Engineering :: Atmospheric Science","Programming Language :: Python :: 3"],author:""},aiida_version:">=1.6.5,<3.0.0",entry_points:{"aiida.data":{"netcdf.data":"aiida_flexpart.data.nc_data:NetCdfData"},"aiida.calculations":{"flexpart.cosmo":{description:["AiiDA calculation plugin wrapping the FLEXPART executable."],spec:{inputs:[{name:"meteo_path",required:!0,valid_types:"List",info:"Path to the folder containing the meteorological input data."},{name:"model_settings",required:!0,valid_types:"",info:""},{name:"outgrid",required:!0,valid_types:"Dict",info:"Input file for the Lagrangian particle dispersion model FLEXPART."},{name:"species",required:!0,valid_types:"RemoteData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"land_use",required:!1,valid_types:"RemoteData",info:"#TODO"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"outgrid_nest",required:!1,valid_types:"Dict, NoneType",info:"Input file for the Lagrangian particle dispersion model FLEXPART. Nested output grid."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_flexpart.calculations.flexpart_cosmo:FlexpartCosmoCalculation"},"flexpart.ifs":{description:["AiiDA calculation plugin wrapping the FLEXPART IFS executable."],spec:{inputs:[{name:"meteo_path",required:!0,valid_types:"List",info:"Path to the folder containing the meteorological input data."},{name:"model_settings",required:!0,valid_types:"",info:""},{name:"outgrid",required:!0,valid_types:"Dict",info:"Input file for the Lagrangian particle dispersion model FLEXPART."},{name:"species",required:!0,valid_types:"RemoteData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"land_use",required:!1,valid_types:"RemoteData",info:"#TODO"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"outgrid_nest",required:!1,valid_types:"Dict, NoneType",info:"Input file for the Lagrangian particle dispersion model FLEXPART. Nested output grid."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Working directory of a previously ran calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_flexpart.calculations.flexpart_ifs:FlexpartIfsCalculation"},"flexpart.post":{description:["AiiDA calculation plugin for post processing."],spec:{inputs:[{name:"input_dir",required:!0,valid_types:"RemoteData",info:"main FLEXPART output dir"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"input_offline_dir",required:!1,valid_types:"RemoteData, NoneType",info:"offline-nested FLEXPART output dir"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_flexpart.calculations.flexpart_post:PostProcessingCalculation"},"collect.sensitivities":{description:["AiiDA calculation plugin for the collection of sensitivities."," The main input are the stashed Netcdf files for the previous FLEXPART"," calculations."],spec:{inputs:[{name:"model",required:!0,valid_types:"str",info:""},{name:"outgrid",required:!0,valid_types:"str",info:""},{name:"outgrid_n",required:!0,valid_types:"bool",info:""},{name:"remote",required:!0,valid_types:"RemoteStashFolderData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"name",required:!1,valid_types:"str, NoneType",info:""},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_flexpart.calculations.collect_sens:CollectSensitivitiesCalculation"},"inversion.calc":{description:["No description available"],spec:{inputs:[{name:"chunk",required:!0,valid_types:"Str",info:"Options are 'year' and 'month'. Default is 'year'"},{name:"chunk_w",required:!0,valid_types:"Str",info:`Width of the individual inversion chunk. These can be wider than + the chunking itself to allow for running average fluxes., + Possible values are 'year' and '3year' for 'chunk.by=year' and, + 'month' and '3month' for 'chunk.by=month'. Default is 'year' + `},{name:"end_date",required:!0,valid_types:"Str",info:"End date (yyyy-mm-dd)"},{name:"inv_params",required:!0,valid_types:"Dict",info:"File containing inversion settings, either as R source file or yaml"},{name:"observations",required:!0,valid_types:"NetCdfData",info:"Dictionary of observations as NetCDF objects"},{name:"remotes",required:!0,valid_types:"NetCdfData",info:"Dictionary of sensitivities as NetCDF objects"},{name:"start_date",required:!0,valid_types:"Str",info:"Start date (yyyy-mm-dd)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_flexpart.calculations.inversion:Inversion"}},"aiida.parsers":{"flexpart.cosmo":"aiida_flexpart.parsers.flexpart_cosmo:FlexpartCosmoParser","flexpart.ifs":"aiida_flexpart.parsers.flexpart_ifs:FlexpartIfsParser","flexpart.post":"aiida_flexpart.parsers.flexpart_post:FlexpartPostParser","collect.sensitivities":"aiida_flexpart.parsers.collect_sens:CollectSensParser","inversion.calc":"aiida_flexpart.parsers.inversion:InvesrionParser"},"aiida.workflows":{"flexpart.multi_dates":"aiida_flexpart.workflows.multi_dates_workflow:FlexpartMultipleDatesWorkflow","flexpart.multi_workflow":"aiida_flexpart.workflows.parent_workflow:ParentWorkflow","inspect.workflow":"aiida_flexpart.workflows.inspect:InspectWorkflow","inversion.workflow":"aiida_flexpart.workflows.inversion_workflow:InversionWorkflow"}},commits_count:47,development_status:"alpha",errors:[],warnings:["W010: Entry point 'netcdf.data' does not start with prefix 'flexpart.'","W010: Entry point 'collect.sensitivities' does not start with prefix 'flexpart.'","W010: Entry point 'inversion.calc' does not start with prefix 'flexpart.'","W010: Entry point 'collect.sensitivities' does not start with prefix 'flexpart.'","W010: Entry point 'inversion.calc' does not start with prefix 'flexpart.'","W010: Entry point 'inspect.workflow' does not start with prefix 'flexpart.'","W010: Entry point 'inversion.workflow' does not start with prefix 'flexpart.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:5},{colorclass:"brown",text:"Parsers",count:5},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:4}],pip_install_cmd:"pip install git+https://github.com/aiidaplugins/aiida-flexpart",is_installable:"True"},"aiida-gaussian":{code_home:"https://github.com/nanotech-empa/aiida-gaussian",entry_point_prefix:"gaussian",pip_url:"aiida-gaussian",plugin_info:"https://raw.githubusercontent.com/nanotech-empa/aiida-gaussian/master/pyproject.toml",name:"aiida-gaussian",package_name:"aiida_gaussian",hosted_on:"github.com",metadata:{release_date:"2024-09-25",description:"AiiDA plugin for the Gaussian quantum chemistry software.",author:"Kristjan Eimre, Pezhman Zarabadi-Poor, Aliaksandr Yakutovich",license:"MIT",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Operating System :: OS Independent","Programming Language :: Python :: 3","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics","Topic :: Software Development :: Libraries :: Python Modules"],version:"2.2.0"},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.calculations":{gaussian:{description:["AiiDA calculation plugin wrapping Gaussian",""," Template:",""," parameters = Dict(dict={"," 'link0_parameters': {"," '%chk':'aiida.chk',"," '%mem': '1024MB',"," '%nprocshared': '2',"," },"," 'functional':'PBE1PBE',"," 'basis_set':'6-31g',"," 'charge': 0,"," 'multiplicity': 1,"," 'route_parameters': {"," 'scf': {'cdiis': None}"," 'nosymm': None,"," 'opt': 'tight',"," },"," })"],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"the folder of a completed gaussian calculation"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"additional input parameters"},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:"Input structure; will be converted to pymatgen object"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The result parameters of the calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"energy_ev",required:!1,valid_types:"Float",info:"Final energy in electronvolts"},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Final optimized structure, if available"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"The retrieved folder data node could not be accessed."},{status:210,message:"The retrieved folder did not contain the output file."},{status:211,message:"The retrieved output log could not be read."},{status:220,message:"The output file could not be parsed."},{status:301,message:"The SCF did not converge and the calculation was terminated."},{status:302,message:"The calculation was terminated due to a logic error in ASyTop."},{status:303,message:"The calculation was terminated due to an inaccurate quadrature in CalDSu."},{status:390,message:"The calculation was terminated due to an error."},{status:391,message:"The log did not contain 'Normal termination' (probably out of time)."}]},class:"aiida_gaussian.calculations:GaussianCalculation"},"gaussian.cubegen":{description:["Plugin to run the cubegen utility",""," Example:",""," parameters = {",' "homo-5": {',' "kind": "AMO=16",',' "npts": -2,'," },",' "spin": {',' "kind": "Spin=SCF",',' "npts": 0,'," },"," }"," Each key corresponds to one produced cube."," key specifies the name of the output node","",' In case of "npts": -1, you have to use the stencil file input:',""," IFlag X0 Y0 Z0 # Output unit number and initial point."," N1 X1 Y1 Z1 # Number of points and step-size in the X-direction."," N2 X2 Y2 Z2 # Number of points and step-size in the Y-direction."," N3 X3 Y3 Z3 # Number of points and step-size in the Z-direction.",""," See more details at https://gaussian.com/cubegen/"],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"dictionary containing entries for cubes to be printed."},{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"the folder of a containing the .fchk"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"gauss_memdef",required:!1,valid_types:"Int, NoneType",info:"Set the GAUSS_MEMDEF env variable to set the max memory in MB."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"retrieve_cubes",required:!1,valid_types:"Bool, NoneType",info:"should the cubes be retrieved?"},{name:"stencil",required:!1,valid_types:"SinglefileData, NoneType",info:"In case of npts=-1, use this cube specification."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"The retrieved folder could not be accessed."},{status:301,message:"The retrieved temporary folder could not be accessed."}]},class:"aiida_gaussian.calculations:CubegenCalculation"},"gaussian.formchk":{description:["Very simple plugin to run the formchk utility"],spec:{inputs:[{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"the folder of a containing the .chk"},{name:"chk_name",required:!1,valid_types:"Str, NoneType",info:"name of the checkpoint file"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"retrieve_fchk",required:!1,valid_types:"Bool, NoneType",info:"retrieve the fchk file"}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_gaussian.calculations:FormchkCalculation"}},"aiida.parsers":{"gaussian.advanced":"aiida_gaussian.parsers.gaussian:GaussianAdvancedParser","gaussian.base":"aiida_gaussian.parsers.gaussian:GaussianBaseParser","gaussian.cubegen_base":"aiida_gaussian.parsers.cubegen:CubegenBaseParser"},"aiida.workflows":{"gaussian.base":{description:["Workchain to run a Gaussian calculation with automated error handling and restarts."],spec:{inputs:[{name:"gaussian",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:350,message:"The calculation failed with an unrecoverable SCF convergence error."},{status:399,message:"The calculation failed with an unrecoverable error."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_gaussian.workchains:GaussianBaseWorkChain"},"gaussian.cubes":{description:["No description available"],spec:{inputs:[{name:"cubegen_code",required:!0,valid_types:"Code",info:""},{name:"formchk_code",required:!0,valid_types:"Code",info:""},{name:"gaussian_calc_folder",required:!0,valid_types:"RemoteData",info:"The gaussian calculation output folder."},{name:"gaussian_output_params",required:!0,valid_types:"Dict",info:"The gaussian calculation output parameters."},{name:"cubegen_parser_name",required:!1,valid_types:"str",info:""},{name:"cubegen_parser_params",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters to cubegen parser."},{name:"dx",required:!1,valid_types:"Float, NoneType",info:"Cube file spacing [ang]."},{name:"edge_space",required:!1,valid_types:"Float, NoneType",info:"Extra cube space in addition to molecule bounding box [ang]."},{name:"generate_density",required:!1,valid_types:"Bool, NoneType",info:"Generate density cube."},{name:"generate_spin_density",required:!1,valid_types:"Bool, NoneType",info:"Generate spin density cube (if applicable)."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"natural_orbitals",required:!1,valid_types:"Bool, NoneType",info:"The cube files are natural orbitals."},{name:"orbital_index_ref",required:!1,valid_types:"Str, NoneType",info:"Reference index, possible choices: 'half_num_el', 'abs'."},{name:"orbital_indexes",required:!1,valid_types:"List, NoneType",info:"Indexes of the orbital cubes to generate."},{name:"retrieve_cubes",required:!1,valid_types:"Bool, NoneType",info:"should the cubes be retrieved?"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_gaussian.workchains:GaussianCubesWorkChain"}}},commits_count:4,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:3},{colorclass:"brown",text:"Parsers",count:3},{colorclass:"green",text:"Workflows",count:2}],pip_install_cmd:"pip install aiida-gaussian",is_installable:"True"},"aiida-gaussian-datatypes":{code_home:"https://github.com/dev-zero/aiida-gaussian-datatypes",documentation_url:"https://github.com/dev-zero/aiida-gaussian-datatypes/blob/master/README.md",entry_point_prefix:"gaussian",pip_url:"aiida-gaussian-datatypes",plugin_info:"https://raw.github.com/dev-zero/aiida-gaussian-datatypes/master/setup.json",name:"aiida-gaussian-datatypes",package_name:"aiida_gaussian_datatypes",hosted_on:"github.com",metadata:{release_date:"2022-07-22",description:"AiiDA data plugin to manage gaussian datatypes (basis sets and pseudopotentials) as first-class citizens",author:"Tiziano Müller",author_email:"tiziano.mueller@chem.uzh.ch",license:"MIT License",home_page:"https://github.com/dev-zero/aiida-gaussian-datatypes",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Operating System :: OS Independent","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Chemistry","Topic :: Software Development :: Libraries :: Python Modules"],version:"0.5.1"},aiida_version:">=1.6.2",entry_points:{"aiida.cmdline.data":{"gaussian.basisset":"aiida_gaussian_datatypes.basisset.cli:cli","gaussian.pseudo":"aiida_gaussian_datatypes.pseudopotential.cli:cli"},"aiida.data":{"gaussian.basisset":"aiida_gaussian_datatypes.basisset.data:BasisSet","gaussian.pseudo":"aiida_gaussian_datatypes.pseudopotential.data:Pseudopotential"},"aiida.groups":{"gaussian.basisset":"aiida_gaussian_datatypes.groups:BasisSetGroup","gaussian.pseudo":"aiida_gaussian_datatypes.groups:PseudopotentialGroup"}},commits_count:0,development_status:"beta",errors:[],warnings:["W009: Prefix 'gaussian' does not follow naming convention."],summaryinfo:[{colorclass:"red",text:"Data",count:2},{colorclass:"orange",text:"Other (Data commands, Groups)",count:4}],pip_install_cmd:"pip install aiida-gaussian-datatypes",is_installable:"True"},"aiida-gollum":{code_home:"https://github.com/garsua/aiida-gollum/",documentation_url:"https://aiida-gollum.readthedocs.io/",entry_point_prefix:"gollum",pip_url:"git+https://github.com/garsua/aiida-gollum",name:"aiida-gollum",package_name:"aiida_gollum",hosted_on:"github.com",metadata:{author:"Victor M. Garcia-Suarez",author_email:"vm.garcia@cinn.es",version:"0.12.0",description:"A plugin for Gollum functionality within AiiDA framework.",classifiers:["License :: OSI Approved :: MIT License","Framework :: AiiDA","Programming Language :: Python :: 2.7","Development Status :: 1 - Alpha"]},aiida_version:">=0.12.0",entry_points:{"aiida.calculations":{"gollum.gollum":"aiida_gollum.calculations.gollum:GollumCalculation"},"aiida.parsers":{"gollum.parser":"aiida_gollum.parsers.gollum:GollumParser"}},commits_count:0,development_status:"planning",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install git+https://github.com/garsua/aiida-gollum"},"aiida-graphql":{code_home:"https://github.com/dev-zero/aiida-graphql",entry_point_prefix:"graphql",pip_url:"aiida-graphql",name:"aiida-graphql",package_name:"aiida_graphql",hosted_on:"github.com",metadata:{release_date:"2019-10-28",description:"Strawberry-based GraphQL API Server for AiiDA",author:"Tiziano Müller",author_email:"tiziano.mueller@chem.uzh.ch",license:"MIT",home_page:"https://github.com/dev-zero/aiida-graphql",classifiers:["Development Status :: 3 - Alpha","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Operating System :: OS Independent","Programming Language :: Python :: 3","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Software Development :: Libraries :: Python Modules"],version:"0.0.2"},aiida_version:">=1.0.0b6,<2.0.0",entry_points:{},commits_count:0,development_status:"alpha",errors:[`E001: Failed to install plugin aiida-graphql
Collecting aiida-graphql
+  Downloading aiida_graphql-0.0.2-py3-none-any.whl.metadata (3.3 kB)
+Collecting aiida<2.0.0,>=1.0.0b6 (from aiida-graphql)
+  Downloading aiida-1.0.1.tar.gz (2.8 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)
+  Downloading strawberry_graphql-0.16.10-py3-none-any.whl.metadata (4.0 kB)
+Requirement already satisfied: aiida-core in /opt/conda/lib/python3.10/site-packages (from aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.5.2)
+Collecting click<8.0,>=7.0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading click-7.1.2-py2.py3-none-any.whl.metadata (2.9 kB)
+Collecting graphql-core<4.0.0,>=3.0.0a0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading graphql_core-3.3.0a6-py3-none-any.whl.metadata (11 kB)
+Collecting hupper<2.0,>=1.5 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading hupper-1.12.1-py3-none-any.whl.metadata (3.7 kB)
+Requirement already satisfied: pygments<3.0,>=2.3 in /opt/conda/lib/python3.10/site-packages (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql) (2.18.0)
+Collecting starlette==0.12.10 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading starlette-0.12.10.tar.gz (46 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 46.3/46.3 kB 3.8 MB/s eta 0:00:00
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting uvicorn==0.10.0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading uvicorn-0.10.0.tar.gz (26 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting h11==0.8.* (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading h11-0.8.1-py2.py3-none-any.whl.metadata (8.0 kB)
+Collecting websockets==8.* (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading websockets-8.1.tar.gz (58 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 58.9/58.9 kB 14.8 MB/s eta 0:00:00
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting httptools==0.0.13 (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading httptools-0.0.13.tar.gz (104 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 104.2/104.2 kB 19.7 MB/s eta 0:00:00
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting uvloop==0.14.0rc2 (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading uvloop-0.14.0rc2.tar.gz (2.0 MB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.0/2.0 MB 12.1 MB/s eta 0:00:00
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Requirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.13.1)
+Requirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.4.2)
+Requirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.8.1)
+Requirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.18.0)
+Requirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.1.10)
+INFO: pip is looking at multiple versions of aiida-core to determine which version is compatible with other requirements. This could take a while.
+Collecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)
+  Downloading strawberry_graphql-0.16.9-py3-none-any.whl.metadata (4.0 kB)
+Collecting uvicorn==0.9.1 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading uvicorn-0.9.1.tar.gz (25 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting uvloop==0.* (from uvicorn==0.9.1->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)
+  Downloading uvloop-0.21.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB)
+Collecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)
+  Downloading strawberry_graphql-0.16.8-py3-none-any.whl.metadata (4.0 kB)
+  Downloading strawberry_graphql-0.16.7-py3-none-any.whl.metadata (4.0 kB)
+Collecting aiida<2.0.0,>=1.0.0b6 (from aiida-graphql)
+  Downloading aiida-1.0.0.tar.gz (1.9 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida to determine which version is compatible with other requirements. This could take a while.
+  Downloading aiida-1.0.0b6.tar.gz (2.6 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting aiida-graphql
+  Downloading aiida_graphql-0.0.1-py3-none-any.whl.metadata (3.4 kB)
+Requirement already satisfied: click~=8.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (8.1.7)
+Requirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.1.0)
+Requirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.16)
+Requirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.20.3)
+Requirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (8.25.0)
+Requirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.18.2)
+Requirement already satisfied: jinja2~=3.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (3.1.4)
+Requirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.7.8)
+Requirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.11.0)
+Requirement already satisfied: numpy~=1.21 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.26.4)
+Requirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.12.0)
+Requirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.21.10)
+Requirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.2.4)
+Requirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (5.9.8)
+Requirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.9.9)
+Requirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.7.3)
+Requirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2021.3)
+Requirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.0.1)
+Requirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.32.3)
+Requirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.0.30)
+Requirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.8.10)
+Requirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (4.66.4)
+Requirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.9.5)
+Requirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.16.0)
+INFO: pip is looking at multiple versions of strawberry-graphql to determine which version is compatible with other requirements. This could take a while.
+Collecting strawberry-graphql<0.17.0,>=0.16.5 (from aiida-graphql)
+  Downloading strawberry_graphql-0.16.6-py3-none-any.whl.metadata (4.0 kB)
+  Downloading strawberry_graphql-0.16.5-py3-none-any.whl.metadata (4.0 kB)
+Collecting aiida-core (from aiida<2.0.0,>=1.0.0b6->aiida-graphql)
+  Downloading aiida_core-2.5.2-py3-none-any.whl.metadata (11 kB)
+INFO: pip is still looking at multiple versions of aiida-core to determine which version is compatible with other requirements. This could take a while.
+Collecting uvicorn==0.9.0 (from strawberry-graphql<0.17.0,>=0.16.5->aiida-graphql)
+  Downloading uvicorn-0.9.0.tar.gz (24 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+ERROR: Cannot install aiida-graphql because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida 1.0.0 depends on aiida-core==1.0
+    aiida 1.0.0b6 depends on aiida-core==1.0.0b6
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W020: Unable to read wheel file from PyPI release:
No entry_points.txt found in wheel
","W003: Missing classifier 'Framework :: AiiDA'"],summaryinfo:[],pip_install_cmd:"pip install aiida-graphql",is_installable:"False"},"aiida-gromacs":{code_home:"https://github.com/PSDI-UK/aiida-gromacs",documentation_url:"https://aiida-gromacs.readthedocs.io/",entry_point_prefix:"gromacs",pip_url:"aiida-gromacs",plugin_info:"https://raw.githubusercontent.com/PSDI-UK/aiida-gromacs/master/pyproject.toml",name:"aiida-gromacs",package_name:"aiida_gromacs",hosted_on:"github.com",metadata:{release_date:"2024-05-23",description:"A plugin for using GROMACS with AiiDA for molecular dymanics simulations.",author_email:"James Gebbie-Rayet , Jas Kalayan ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"2.0.8"},aiida_version:">=2.4.0,<3",entry_points:{"aiida.calculations":{"gromacs.editconf":{description:["AiiDA calculation plugin wrapping the 'gmx editconf' executable.",""," AiiDA plugin wrapper for adding a simulation box to structure file."],spec:{inputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Input structure file."},{name:"parameters",required:!0,valid_types:"EditconfParameters",info:"Command line parameters for gmx editconf."},{name:"bf_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Generic data file."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"n_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Index file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Output file containing simulation box."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"mead_file",required:!1,valid_types:"SinglefileData",info:"Coordination file for MEAD"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.editconf:EditconfCalculation"},"gromacs.genericMD":{description:["AiiDA calculation plugin wrapping an executable with user defined"," input and output files."],spec:{inputs:[{name:"code",required:!0,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"input_files",required:!1,valid_types:"SinglefileData",info:"Dictionary of input files."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"output_files",required:!1,valid_types:"List, NoneType",info:"List of output file names."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."},{status:301,message:"Specified output file not produced by command."}]},class:"aiida_gromacs.calculations.genericMD:GenericCalculation"},"gromacs.genion":{description:["AiiDA calculation plugin wrapping the 'gmx genion' executable.",""," AiiDA plugin wrapper for converting PDB files to GRO files."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"GenionParameters",info:"Command line parameters for gmx genion"},{name:"topfile",required:!0,valid_types:"SinglefileData",info:"Input topology file."},{name:"tprfile",required:!0,valid_types:"SinglefileData",info:"Input tpr file."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"instructions_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Instructions for generating index file"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"n_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Index file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Output gro file with ions added."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"topfile",required:!0,valid_types:"SinglefileData",info:"Output topology with ions added."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.genion:GenionCalculation"},"gromacs.grompp":{description:["AiiDA calculation plugin wrapping the 'gmx grompp' executable.",""," AiiDA plugin wrapper for converting PDB files to GRO files."],spec:{inputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Input structure"},{name:"mdpfile",required:!0,valid_types:"SinglefileData",info:"grompp run file."},{name:"parameters",required:!0,valid_types:"GromppParameters",info:"Command line parameters for gmx grompp"},{name:"topfile",required:!0,valid_types:"SinglefileData",info:"Input topology"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"e_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Energy file"},{name:"itp_dirs",required:!1,valid_types:"FolderData",info:"Forcefield descriptions"},{name:"itp_files",required:!1,valid_types:"SinglefileData",info:"Restraint files"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"n_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Index file"},{name:"qmi_file",required:!1,valid_types:"SinglefileData, NoneType",info:"QM input file"},{name:"r_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Structure file"},{name:"rb_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Structure file"},{name:"ref_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Full precision trajectory file"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"t_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Full precision trajectory file"}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"tprfile",required:!0,valid_types:"SinglefileData",info:"Output gro file ready for adding ions."},{name:"imd_file",required:!1,valid_types:"SinglefileData",info:"Coordinate file in Gromos-87 format"},{name:"po_file",required:!1,valid_types:"SinglefileData",info:"grompp input file with MD parameters"},{name:"pp_file",required:!1,valid_types:"SinglefileData",info:"Topology file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.grompp:GromppCalculation"},"gromacs.make_ndx":{description:["AiiDA calculation plugin wrapping the 'gmx make_ndx' executable.",""," AiiDA plugin wrapper for converting PDB files to GRO files."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Make_ndxParameters",info:"Command line parameters for gmx make_ndx"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"grofile",required:!1,valid_types:"SinglefileData, NoneType",info:"Structure file: gro g96 pdb brk ent esp tpr"},{name:"instructions_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Instructions for generating index file"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"n_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Index file"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"n_file_out",required:!0,valid_types:"SinglefileData",info:"Index file"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.make_ndx:Make_ndxCalculation"},"gromacs.mdrun":{description:["AiiDA calculation plugin wrapping the 'gmx mdrun' executable.",""," AiiDA plugin wrapper for converting PDB files to GRO files."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"MdrunParameters",info:"Command line parameters for gmx mdrun"},{name:"tprfile",required:!0,valid_types:"SinglefileData",info:"Input structure."},{name:"awh_file",required:!1,valid_types:"SinglefileData, NoneType",info:"xvgr/xmgr file"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"cpi_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Checkpoint file"},{name:"ei_file",required:!1,valid_types:"SinglefileData, NoneType",info:"ED sampling input"},{name:"membed_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Generic data file"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"mn_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Index file"},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"mp_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Topology file"},{name:"multidir_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Run directory"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"rerun_file",required:!1,valid_types:"SinglefileData, NoneType",info:"Trajectory: xtc trr cpt gro g96 pdb tng"},{name:"table_file",required:!1,valid_types:"SinglefileData, NoneType",info:"xvgr/xmgr file"},{name:"tableb_file",required:!1,valid_types:"SinglefileData, NoneType",info:"xvgr/xmgr file"},{name:"tablep_file",required:!1,valid_types:"SinglefileData, NoneType",info:"xvgr/xmgr file"}],outputs:[{name:"enfile",required:!0,valid_types:"SinglefileData",info:"Output energy file."},{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Output structure file."},{name:"logfile",required:!0,valid_types:"SinglefileData",info:"Output log file."},{name:"logfile_metadata",required:!0,valid_types:"Dict",info:"metadata exracted from gromacs logfile"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"trrfile",required:!0,valid_types:"SinglefileData",info:"Output trajectory."},{name:"cpo_file",required:!1,valid_types:"SinglefileData",info:"Checkpoint file."},{name:"dhdl_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"eo_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"field_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"if_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"mtx_file",required:!1,valid_types:"SinglefileData",info:"Hessian Matrix"},{name:"pf_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"px_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"ra_file",required:!1,valid_types:"SinglefileData",info:"Log file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"ro_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"rs_file",required:!1,valid_types:"SinglefileData",info:"Log file"},{name:"rt_file",required:!1,valid_types:"SinglefileData",info:"Log file"},{name:"swap_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"tpi_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"tpid_file",required:!1,valid_types:"SinglefileData",info:"xvgr/xmgr file"},{name:"x_file",required:!1,valid_types:"SinglefileData",info:"Compressed trajectory (tng format or portable xdr format)"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.mdrun:MdrunCalculation"},"gromacs.pdb2gmx":{description:["AiiDA calculation plugin wrapping the 'gmx pdb2gmx' executable.",""," AiiDA plugin wrapper for converting PDB files to GRO files."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Pdb2gmxParameters",info:"Command line parameters for gmx pdb2gmx"},{name:"pdbfile",required:!0,valid_types:"SinglefileData",info:"Input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Output forcefield compliant file."},{name:"itpfile",required:!0,valid_types:"SinglefileData",info:"Output forcefield compliant file."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"topfile",required:!0,valid_types:"SinglefileData",info:"Output forcefield compliant file."},{name:"n_file",required:!1,valid_types:"SinglefileData",info:"Output index file"},{name:"q_file",required:!1,valid_types:"SinglefileData",info:"Output Structure file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.pdb2gmx:Pdb2gmxCalculation"},"gromacs.solvate":{description:["AiiDA calculation plugin wrapping the 'gmx solvate' executable.",""," AiiDA plugin wrapper for solvating a molecular system."],spec:{inputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Input structure"},{name:"parameters",required:!0,valid_types:"SolvateParameters",info:"Command line parameters for gmx solvate."},{name:"topfile",required:!0,valid_types:"SinglefileData",info:"Input topology"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"command",required:!1,valid_types:"Str, NoneType",info:"The command used to execute the job."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"grofile",required:!0,valid_types:"SinglefileData",info:"Output solvated gro file."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stdout",required:!0,valid_types:"SinglefileData",info:"stdout"},{name:"topfile",required:!0,valid_types:"SinglefileData",info:"Output topology file."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."}]},class:"aiida_gromacs.calculations.solvate:SolvateCalculation"}},"aiida.cmdline.data":{provenance:"aiida_gromacs.commands.provenance:provenance"},"aiida.data":{"gromacs.editconf":"aiida_gromacs.data.editconf:EditconfParameters","gromacs.genion":"aiida_gromacs.data.genion:GenionParameters","gromacs.grompp":"aiida_gromacs.data.grompp:GromppParameters","gromacs.make_ndx":"aiida_gromacs.data.make_ndx:Make_ndxParameters","gromacs.mdrun":"aiida_gromacs.data.mdrun:MdrunParameters","gromacs.pdb2gmx":"aiida_gromacs.data.pdb2gmx:Pdb2gmxParameters","gromacs.solvate":"aiida_gromacs.data.solvate:SolvateParameters"},"aiida.parsers":{"gromacs.editconf":"aiida_gromacs.parsers.editconf:EditconfParser","gromacs.genericMD":"aiida_gromacs.parsers.genericMD:GenericParser","gromacs.genion":"aiida_gromacs.parsers.genion:GenionParser","gromacs.grompp":"aiida_gromacs.parsers.grompp:GromppParser","gromacs.make_ndx":"aiida_gromacs.parsers.make_ndx:Make_ndxParser","gromacs.mdrun":"aiida_gromacs.parsers.mdrun:MdrunParser","gromacs.pdb2gmx":"aiida_gromacs.parsers.pdb2gmx:Pdb2gmxParser","gromacs.solvate":"aiida_gromacs.parsers.solvate:SolvateParser"},"aiida.workflows":{"gromacs.setup":{description:["WorkChain for setting up a gromacs simulation automatically."],spec:{inputs:[{name:"editconfparameters",required:!0,valid_types:"EditconfParameters",info:"Command line parameters for gmx editconf"},{name:"genionparameters",required:!0,valid_types:"GenionParameters",info:"Command line parameters for gmx genion"},{name:"gromppionsparameters",required:!0,valid_types:"GromppParameters",info:"Command line parameters for gmx grompp"},{name:"gromppminparameters",required:!0,valid_types:"GromppParameters",info:"Command line parameters for gmx grompp minimisation run"},{name:"gromppnptparameters",required:!0,valid_types:"GromppParameters",info:"Command line parameters for gmx grompp npt equilibration run"},{name:"gromppnvtparameters",required:!0,valid_types:"GromppParameters",info:"Command line parameters for gmx grompp nvt equilibration run"},{name:"gromppprodparameters",required:!0,valid_types:"GromppParameters",info:"Command line parameters for gmx grompp production run"},{name:"ionsmdp",required:!0,valid_types:"SinglefileData",info:"MD parameters for adding ions."},{name:"local_code",required:!0,valid_types:"Code",info:""},{name:"mdrunparameters",required:!0,valid_types:"MdrunParameters",info:"Command line parameters for gmx mdrun production run"},{name:"minimiseparameters",required:!0,valid_types:"MdrunParameters",info:"Command line parameters for gmx mdrun minimisation run"},{name:"minmdp",required:!0,valid_types:"SinglefileData",info:"MD parameters for minimisation."},{name:"nptmdp",required:!0,valid_types:"SinglefileData",info:"MD parameters for NPT equilibration."},{name:"nptparameters",required:!0,valid_types:"MdrunParameters",info:"Command line parameters for gmx mdrun npt equilibration run"},{name:"nvtmdp",required:!0,valid_types:"SinglefileData",info:"MD parameters for NVT equilibration."},{name:"nvtparameters",required:!0,valid_types:"MdrunParameters",info:"Command line parameters for gmx mdrun nvt equilibration run"},{name:"pdb2gmxparameters",required:!0,valid_types:"Pdb2gmxParameters",info:"Command line parameters for gmx pdb2gmx"},{name:"pdbfile",required:!0,valid_types:"SinglefileData",info:"Input structure."},{name:"prodmdp",required:!0,valid_types:"SinglefileData",info:"MD parameters for production run."},{name:"solvateparameters",required:!0,valid_types:"SolvateParameters",info:"Command line parameters for gmx solvate"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"remote_code",required:!1,valid_types:"Code, NoneType",info:""}],outputs:[{name:"result",required:!0,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_gromacs.workflows.simsetup:SetupWorkChain"}},console_scripts:{createarchive:"aiida_gromacs.cli.createarchive:cli",genericMD:"aiida_gromacs.cli.genericMD:cli",gmx_editconf:"aiida_gromacs.cli.editconf:cli",gmx_genion:"aiida_gromacs.cli.genion:cli",gmx_grompp:"aiida_gromacs.cli.grompp:cli",gmx_make_ndx:"aiida_gromacs.cli.make_ndx:cli",gmx_mdrun:"aiida_gromacs.cli.mdrun:cli",gmx_pdb2gmx:"aiida_gromacs.cli.pdb2gmx:cli",gmx_solvate:"aiida_gromacs.cli.solvate:cli"}},commits_count:174,development_status:"stable",errors:[],warnings:["W010: Entry point 'provenance' does not start with prefix 'gromacs.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:8},{colorclass:"brown",text:"Parsers",count:8},{colorclass:"red",text:"Data",count:7},{colorclass:"green",text:"Workflows",count:1},{colorclass:"purple",text:"Console scripts",count:9},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install aiida-gromacs",is_installable:"True"},"aiida-grouppathx":{code_home:"https://github.com/zhubonan/aiida-grouppathx",development_status:"beta",entry_point_prefix:"grouppathx",pip_url:"aiida-grouppathx",plugin_info:"https://raw.githubusercontent.com/zhubonan/aiida-grouppathx/master/pyproject.toml",name:"aiida-grouppathx",package_name:"aiida_grouppathx",hosted_on:"github.com",metadata:{release_date:"2024-11-10",description:"AiiDA plugin provides the GroupPathX class",author_email:"Bonan Zhu ",classifiers:["Development Status :: 3 - Alpha","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.2.3"},aiida_version:">=1.6.4,<3",entry_points:{"aiida.cmdline.data":{gpx:"aiida_grouppathx.cli:grouppathx_cli"}},commits_count:13,errors:[],warnings:["W005: Development status in classifiers (alpha) does not match development_status in metadata (beta)","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'gpx' does not start with prefix 'grouppathx.'"],summaryinfo:[{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install aiida-grouppathx",is_installable:"True"},"aiida-gudhi":{code_home:"https://github.com/ltalirz/aiida-gudhi",development_status:"beta",entry_point_prefix:"gudhi",pip_url:"aiida-gudhi",plugin_info:"https://raw.github.com/ltalirz/aiida-gudhi/master/setup.json",name:"aiida-gudhi",package_name:"aiida_gudhi",hosted_on:"github.com",metadata:{release_date:"2018-06-21",description:"AiiDA plugin for the [GUDHI](http://gudhi.gforge.inria.fr/) library for topological data analysis.",author:"Leopold Talirz",author_email:"leopold.talirz@gmail.com",license:"MIT",home_page:"https://github.com/ltalirz/aiida-gudhi",classifiers:["Programming Language :: Python"],version:"0.1.0a3"},aiida_version:"*",entry_points:{"aiida.calculations":{"gudhi.rdm":"aiida_gudhi.calculations.rips:RipsDistanceMatrixCalculation"},"aiida.data":{"gudhi.rdm":"aiida_gudhi.data.rips:RipsDistanceMatrixParameters"},"aiida.parsers":{"gudhi.rdm":"aiida_gudhi.parsers.rips:RipsParser"}},commits_count:0,errors:[],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1}],pip_install_cmd:"pip install --pre aiida-gudhi",is_installable:"True"},"aiida-gulp":{code_home:"https://github.com/aiidaplugins/aiida-gulp",development_status:"beta",documentation_url:"https://aiida-gulp.readthedocs.io",entry_point_prefix:"gulp",pip_url:"aiida-gulp",plugin_info:"https://raw.githubusercontent.com/aiidaplugins/aiida-gulp/master/setup.json",name:"aiida-gulp",package_name:"aiida_gulp",hosted_on:"github.com",metadata:{release_date:"2019-10-30",description:"AiiDA plugin for running the GULP MD code",author:"Chris Sewell",author_email:"chrisj_sewell@hotmail.com",license:"MIT",home_page:"https://github.com/chrisjsewell/aiida-gulp",classifiers:["Framework :: AiiDA","Programming Language :: Python","Programming Language :: Python :: 2.7","Programming Language :: Python :: 3.6","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics"],version:"0.10.0b5"},aiida_version:"1.0.0b5",entry_points:{"aiida.calculations":{"gulp.fitting":"aiida_gulp.calculations.gulp_fitting:GulpFittingCalculation","gulp.optimize":"aiida_gulp.calculations.gulp_optimize:GulpOptCalculation","gulp.single":"aiida_gulp.calculations.gulp_single:GulpSingleCalculation"},"aiida.cmdline.data":{"gulp.potentials":"aiida_gulp.cmndline.potentials:potentials"},"aiida.data":{"gulp.potential":"aiida_gulp.data.potential:EmpiricalPotential","gulp.symmetry":"aiida_gulp.data.symmetry:SymmetryData"},"aiida.parsers":{"gulp.fitting":"aiida_gulp.parsers.parse_fitting:GulpFittingParser","gulp.optimize":"aiida_gulp.parsers.parse_opt:GulpOptParser","gulp.single":"aiida_gulp.parsers.parse_single:GulpSingleParser"},"aiida.workflows":{},console_scripts:{gulp_mock:"aiida_gulp.tests.mock_gulp:main"},"gulp.potentials":{lj:"aiida_gulp.potentials.lj:PotentialWriterLJ",reaxff:"aiida_gulp.potentials.reaxff:PotentialWriterReaxff"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-gulp
Collecting aiida-gulp
+  Downloading aiida_gulp-0.10.0b5-py3-none-any.whl.metadata (5.4 kB)
+INFO: pip is looking at multiple versions of aiida-gulp to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-gulp==0.10.0b5 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-gulp 0.10.0b5 depends on aiida-core==1.0.0b5
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:3},{colorclass:"brown",text:"Parsers",count:3},{colorclass:"red",text:"Data",count:2},{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Data commands, Gulp potentials)",count:3}],pip_install_cmd:"pip install --pre aiida-gulp",is_installable:"False"},"aiida-icon":{entry_point_prefix:"icon",pip_url:"aiida-icon",documentation_url:"https://aiida-icon.github.io/aiida-icon/",code_home:"https://github.com/aiida-icon/aiida-icon",plugin_info:"https://raw.githubusercontent.com/aiida-icon/aiida-icon/main/pyproject.toml",name:"aiida-icon",package_name:"aiida_icon",hosted_on:"github.com",metadata:{release_date:"2024-11-11",description:"AiiDA Plugin to run simulations with the ICON weather & climate model",author_email:"Rico Häuselmann ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.13","Programming Language :: Python :: Implementation :: CPython"],version:"0.4.0"},aiida_version:">=2.5",entry_points:{"aiida.calculations":{"icon.icon":{description:["AiiDA calculation to run ICON."],spec:{inputs:[{name:"cloud_opt_props",required:!0,valid_types:"RemoteData",info:""},{name:"dmin_wetgrowth_lookup",required:!0,valid_types:"RemoteData",info:""},{name:"dynamics_grid_file",required:!0,valid_types:"RemoteData",info:""},{name:"ecrad_data",required:!0,valid_types:"RemoteData",info:""},{name:"master_namelist",required:!0,valid_types:"SinglefileData",info:""},{name:"model_namelist",required:!0,valid_types:"SinglefileData",info:""},{name:"rrtmg_sw",required:!0,valid_types:"RemoteData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_file",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"wrapper_script",required:!1,valid_types:"SinglefileData, NoneType",info:""}],outputs:[{name:"all_restart_files",required:!0,valid_types:"",info:""},{name:"finish_status",required:!0,valid_types:"",info:""},{name:"latest_restart_file",required:!0,valid_types:"",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"ICON did not create a restart file or directory."},{status:301,message:"Could not read the finish.status file."},{status:302,message:"The final status was not 'OK or RESTART', check the finish_status output."},{status:303,message:"Could not fully parse due to an unsupported feature, please check the log."},{status:304,message:"Some outputs might be missing, check the log for explanations."},{status:310,message:"ICON was expected to produce a restart file but did not."}]},class:"aiida_icon.calculations:IconCalculation"}},"aiida.parsers":{"icon.icon":"aiida_icon.calculations:IconParser"}},commits_count:22,development_status:"beta",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install aiida-icon",is_installable:"True"},"aiida-inq":{entry_point_prefix:"inq",plugin_info:"https://github.com/LLNL/aiida-inq/blob/main/pyproject.toml",code_home:"https://github.com/LLNL/aiida-inq",pip_url:"aiida-inq",name:"aiida-inq",package_name:"aiida_inq",hosted_on:"github.com",metadata:{release_date:"2024-10-08",description:"The official AiiDA plugin for INQ.",author_email:"Nathan Keilbart ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Topic :: Scientific/Engineering"],version:"0.1.0"},aiida_version:">=2.0,<3.0",entry_points:{"aiida.calculations":{"inq.create_kpoints_from_distance":{description:["Generate a kpoint mesh for a given structure.",""," :param structure: StructureData to which the mesh will be applied."," :param kspacing: Spacing between kpoints in reciprocal space.",""," :returns: KpointsData object with mesh."],spec:{inputs:[{name:"kspacing",required:!0,valid_types:"Data",info:"Spacing between kpoints in reciprocal space."},{name:"structure",required:!0,valid_types:"Data",info:"StructureData to which the mesh will be applied."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_inq.calculations.functions.create_kpoints_from_distance:create_kpoints_from_distance"},"inq.inq":{description:["Base calculation class for the INQ code."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters for the input file."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Optional working directory of a previous calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job is performed."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:201,message:"At minimum the energy cutoff must be specified."},{status:202,message:"No run type was specified in the input parameters."},{status:203,message:"Input parameter was not specified for this result."}]},class:"aiida_inq.calculations.inq:InqCalculation"}},"aiida.parsers":{"inq.inq":"aiida_inq.parsers.inq:InqParser"},"aiida.workflows":{"inq.base":{description:["Workchain to run an Inq calculation with automated error handling "," and restarts."],spec:{inputs:[{name:"inq",required:!0,valid_types:"Data",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"Whether to clean all related work folders."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Kpoint grid."},{name:"kpoints_spacing",required:!1,valid_types:"Float, NoneType",info:"The spacing between kpoints in reciprocal space."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_inq.workflows.base:InqBaseWorkChain"},"inq.convergence":{description:["Workchain to run convergence tests using the Inq calculator."],spec:{inputs:[{name:"clean_workdir",required:!0,valid_types:"Bool",info:"If `True`, work directories of all called calculations will be cleaned at the end of the workflow."},{name:"conv",required:!0,valid_types:"Data",info:"Inputs for the INQ Base Workchain."},{name:"structure",required:!0,valid_types:"StructureData",info:"The starting structure"},{name:"energy_delta",required:!1,valid_types:"Float, NoneType",info:"The value used to check if the total energy has converged. Since the parser returns values in eV, make sure to scale the value accordingly."},{name:"energy_start",required:!1,valid_types:"Int, NoneType",info:"If provided, will use this energy cutoff as a starting point. Otherwise, the suggested energy cutoffs will be used from the pseudos.yaml protocol file. Units are considered to be Ha."},{name:"energy_step",required:!1,valid_types:"Int, NoneType",info:"Default value for increasing the energy cutoff value. Units considered to be in Ha."},{name:"kspacing_start",required:!1,valid_types:"Float, NoneType",info:"Starting kspacing value for convergence testing."},{name:"kspacing_step",required:!1,valid_types:"Float, NoneType",info:"Step value for reducing kspacing value."},{name:"max_iter",required:!1,valid_types:"Int, NoneType",info:"Maximum number of iterations to perform for both energy and kspacing calculations."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"suggested",required:!0,valid_types:"Dict",info:"Suggested values for energy cutoff and kspacing."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"An INQ calculation failed."},{status:402,message:"Reached the maximum number of iterations for the workchain."}]},class:"aiida_inq.workflows.convergence:InqConvergenceWorkChain"},"inq.tddft":{description:["Workchain to run convergence tests using the Inq calculator."],spec:{inputs:[{name:"clean_workdir",required:!0,valid_types:"Bool",info:"If `True`, work directories of all called calculations will be cleaned at the end of the workflow."},{name:"gs",required:!0,valid_types:"Data",info:"Inputs for the Ground State calculation."},{name:"structure",required:!0,valid_types:"StructureData",info:"The starting structure."},{name:"tddft",required:!0,valid_types:"Data",info:"Inputs for the TDDFT calculation."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"An INQ calculation failed."}]},class:"aiida_inq.workflows.tddft:InqTDDFTWorkChain"}}},commits_count:77,development_status:"beta",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:3}],pip_install_cmd:"pip install aiida-inq",is_installable:"True"},"aiida-kkr":{code_home:"https://github.com/JuDFTteam/aiida-kkr/tree/develop",development_status:"stable",documentation_url:"https://aiida-kkr.readthedocs.io/",entry_point_prefix:"kkr",pip_url:"aiida-kkr",plugin_info:"https://raw.githubusercontent.com/JuDFTteam/aiida-kkr/master/pyproject.toml",name:"aiida-kkr",package_name:"aiida_kkr",hosted_on:"github.com",metadata:{release_date:"2024-11-21",description:"AiiDA plugin for the JuKKR codes",author_email:"Philipp Ruessmann , Jens Broeder , Fabian Bertoldo ",classifiers:["Development Status :: 4 - Beta","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Physics"],version:"2.2.1"},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.calculations":{"kkr.kkr":{description:["AiiDA calculation plugin for a KKR calculation."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Use a node that specifies the input parameters"},{name:"parent_folder",required:!0,valid_types:"RemoteData",info:` +Use a remote or local repository folder as parent folder +(also for restarts and similar). It should contain all the needed +files for a KKR calc, only edited files should be uploaded from the +repository. +`},{name:"anomalous_density",required:!1,valid_types:"FolderData, NoneType",info:` +FolderData that contains anomalous density input files for +the KKRhost BdG calculation. If these are not give the code looks +for them in the retrieved of the parent calculation and takes them +from there.`},{name:"bfield",required:!1,valid_types:"Dict, NoneType",info:` +Non-collinear exteral B-field used for constraint calculations. + +The Dict node should be of the form +initial_noco_angles = Dict(dict={ + 'theta': [theta_at1, theta_at2, ..., theta_atN], + # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], + # list phi values in degrees (0..360) + 'magnitude': [magnitude at_1, ..., magnitude at_N] + # list of magnitude of the applied fields in Ry units +}) +Note: The length of the theta, phi and magnitude lists have to be +equal to the number of atoms. +`},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"deciout_parent",required:!1,valid_types:"RemoteData, NoneType",info:"KkrCalculation RemoteData folder from deci-out calculation"},{name:"impurity_info",required:!1,valid_types:"Dict, NoneType",info:` +Use a Parameter node that specifies properties for a following +impurity calculation (e.g. setting of impurity cluster in scoef +file that is automatically created).`},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` +Initial non-collinear angles for the magnetic moments of +the impurities. These values will be written into the +\`kkrflex_angle\` input file of KKRimp. +The Dict node should be of the form +initial_noco_angles = Dict(dict={ + 'theta': [theta_at1, theta_at2, ..., theta_atN], + # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], + # list phi values in degrees (0..360) + 'fix_dir': [True/False at_1, ..., True/False at_N] + # list of booleans indicating if the direction of the magnetic + # moment should be fixed or is allowed relax (True means keep the + # direction of the magnetic moment fixed) +}) +Note: The length of the theta, phi and fix_dir lists have to be +equal to the number of atoms. +`},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:` +Use a KpointsData node that specifies the kpoints for which a +bandstructure (i.e. 'qdos') calculation should be performed.`},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"retrieve_kkrflex",required:!1,valid_types:"Bool, NoneType",info:` +For a GF writeout calculation, determine whether or not +the kkrflex_* files are copied to the retrieved (can clutter the +database) or are ony left in the remote folder. +`},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:` +Settings for running a LDA+U calculation. The Dict node should be of the form + settings_LDAU = Dict(dict={'iatom=0':{ + 'L': 3, # l-block which gets U correction (1: p, 2: d, 3: f-electrons) + 'U': 7., # U value in eV + 'J': 0.75, # J value in eV + 'Eref_EF': 0., # reference energy in eV relative to the Fermi energy. This is the energy where the projector wavefunctions are calculated (should be close in energy where the states that are shifted lie (e.g. for Eu use the Fermi energy)) + }}) + Note: you can add multiple entries like the one for iatom==0 in this example. The atom index refers to the corresponding atom in the impurity cluster. +`}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"results of the KKR calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"KKR output file not found"},{status:302,message:"KKR parser retuned an error"},{status:303,message:"Could not find shapefun from voronoi parent"}]},class:"aiida_kkr.calculations.kkr:KkrCalculation"},"kkr.kkrimp":{description:["AiiDA calculation plugin for a KKRimp calculation."],spec:{inputs:[{name:"host_Greenfunction_folder",required:!0,valid_types:"RemoteData",info:"Use a node that specifies the host KKR calculation contaning the host Green function and tmatrix (KkrCalculation with impurity_info input)."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"host_Greenfunction_folder_Efshift",required:!1,valid_types:"RemoteData, NoneType",info:"Use a node that specifies the host KKR calculation contaning the host Green function and tmatrix with Fermi level shift (used to set Fermi level)."},{name:"impurity_info",required:!1,valid_types:"Dict, NoneType",info:"Use a parameter node that specifies properties for a immpurity calculation."},{name:"impurity_potential",required:!1,valid_types:"SinglefileData, NoneType",info:"Use a node that contains the input potential."},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` +Initial non-collinear angles for the magnetic moments of the impurities. These values will be written into the \`kkrflex_angle\` input file of KKRimp. +The Dict node should be of the form + initial_noco_angles = Dict({ + 'theta': [theta_at1, theta_at2, ..., theta_atN], # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], # list phi values in degrees (0..360) + 'fix_dir': [True, False, ..., True/False], # list of booleans indicating of the direction of the magentic moment should be fixed or is allowed to be updated (True means keep the direction of the magnetic moment fixed) + }) + +Note: The length of the theta, phi and fix_dir lists have to be equal to the number of atoms in the impurity cluster. +`},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Use a node that specifies the input parameters (calculation settings)."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a node that specifies a parent KKRimp calculation."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"rimpshift",required:!1,valid_types:"Dict, NoneType",info:` +Shift for atoms in the impurity cluster used in U-transformation. + +The Dict node should be of the form + rimpshift = Dict({'shifts': [[0., 0., 0.], ... ]}) +where the shifts are given in atomic units (i.e. the internal KKR units). + +Note: The length of the 'shifts' attribute should be an array with three numbers indicating the shift for each atom in the impurity cluster. +`},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:` +Settings for running a LDA+U calculation. The Dict node should be of the form + settings_LDAU = Dict({'iatom=0':{ + 'L': 3, # l-block which gets U correction (1: p, 2: d, 3: f-electrons) + 'U': 7., # U value in eV + 'J': 0.75, # J value in eV + 'Eref_EF': 0., # reference energy in eV relative to the Fermi energy. This is the energy where the projector wavefunctions are calculated (should be close in energy where the states that are shifted lie (e.g. for Eu use the Fermi energy)) + }}) + +Note: you can add multiple entries like the one for iatom==0 in this example. The atom index refers to the corresponding atom in the impurity cluster. +`}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"results of the KKRimp calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"Retrieved folder of KKRimp calculation not found."},{status:302,message:"KKRimp parser returned an error."}]},class:"aiida_kkr.calculations.kkrimp:KkrimpCalculation"},"kkr.kkrimporter":{description:["Importer dummy calculation for a previous KKR run",""," :param remote_workdir: Absolute path to the directory where the job was run."," The transport of the computer you link ask input to the calculation is"," the transport that will be used to retrieve the calculation's files."," Therefore, ``remote_workdir`` should be the absolute path to the job's"," directory on that computer."," :type remote_workdir: str"," :param input_file_names: The file names of the job's input file."," :type input_file_name: dict with str entries"," :param output_file_name: The file names of the job's output file (i.e. the"," file containing the stdout of kkr.x)."," :type output_file_name: dict with str entries"],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Use a node that specifies the input parameters"},{name:"parent_folder",required:!0,valid_types:"RemoteData",info:` +Use a remote or local repository folder as parent folder +(also for restarts and similar). It should contain all the needed +files for a KKR calc, only edited files should be uploaded from the +repository. +`},{name:"anomalous_density",required:!1,valid_types:"FolderData, NoneType",info:` +FolderData that contains anomalous density input files for +the KKRhost BdG calculation. If these are not give the code looks +for them in the retrieved of the parent calculation and takes them +from there.`},{name:"bfield",required:!1,valid_types:"Dict, NoneType",info:` +Non-collinear exteral B-field used for constraint calculations. + +The Dict node should be of the form +initial_noco_angles = Dict(dict={ + 'theta': [theta_at1, theta_at2, ..., theta_atN], + # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], + # list phi values in degrees (0..360) + 'magnitude': [magnitude at_1, ..., magnitude at_N] + # list of magnitude of the applied fields in Ry units +}) +Note: The length of the theta, phi and magnitude lists have to be +equal to the number of atoms. +`},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"deciout_parent",required:!1,valid_types:"RemoteData, NoneType",info:"KkrCalculation RemoteData folder from deci-out calculation"},{name:"impurity_info",required:!1,valid_types:"Dict, NoneType",info:` +Use a Parameter node that specifies properties for a following +impurity calculation (e.g. setting of impurity cluster in scoef +file that is automatically created).`},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` +Initial non-collinear angles for the magnetic moments of +the impurities. These values will be written into the +\`kkrflex_angle\` input file of KKRimp. +The Dict node should be of the form +initial_noco_angles = Dict(dict={ + 'theta': [theta_at1, theta_at2, ..., theta_atN], + # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], + # list phi values in degrees (0..360) + 'fix_dir': [True/False at_1, ..., True/False at_N] + # list of booleans indicating if the direction of the magnetic + # moment should be fixed or is allowed relax (True means keep the + # direction of the magnetic moment fixed) +}) +Note: The length of the theta, phi and fix_dir lists have to be +equal to the number of atoms. +`},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:` +Use a KpointsData node that specifies the kpoints for which a +bandstructure (i.e. 'qdos') calculation should be performed.`},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"retrieve_kkrflex",required:!1,valid_types:"Bool, NoneType",info:` +For a GF writeout calculation, determine whether or not +the kkrflex_* files are copied to the retrieved (can clutter the +database) or are ony left in the remote folder. +`},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:` +Settings for running a LDA+U calculation. The Dict node should be of the form + settings_LDAU = Dict(dict={'iatom=0':{ + 'L': 3, # l-block which gets U correction (1: p, 2: d, 3: f-electrons) + 'U': 7., # U value in eV + 'J': 0.75, # J value in eV + 'Eref_EF': 0., # reference energy in eV relative to the Fermi energy. This is the energy where the projector wavefunctions are calculated (should be close in energy where the states that are shifted lie (e.g. for Eu use the Fermi energy)) + }}) + Note: you can add multiple entries like the one for iatom==0 in this example. The atom index refers to the corresponding atom in the impurity cluster. +`}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"results of the KKR calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"KKR output file not found"},{status:302,message:"KKR parser retuned an error"},{status:303,message:"Could not find shapefun from voronoi parent"}]},class:"aiida_kkr.calculations.kkrimporter:KkrImporterCalculation"},"kkr.kkrnano":{description:["AiiDA calculation plugin for a KKRnano calculation"],spec:{inputs:[{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"convert",required:!1,valid_types:"Bool, NoneType",info:"Activate to use together with set up convert code in order to retrieve potential files."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"nocoangles",required:!1,valid_types:"Dict, NoneType",info:"Dict node that specifies the starting angles for non-colinear calculations (only needed in conjunction with non-colinear calculations, i. e. KORBIT=1 (which is also necessary for SOC calculations!))"},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Dict node that specifies the input parameters for KKRnano (k-point density etc.)"},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Use a node that specifies a parent KKRnano or voronoi calculation"},{name:"passed_lattice_param_angs",required:!1,valid_types:"Float, NoneType",info:"Use a prespecified lattice constant in Angstrom as input for KKRnano, i. e. in the input.conf file. Default is the length of the longest Bravais vector in the structure object used for the voronoi calculation. This can be useful in the context of treating supercells."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"strucwithpot",required:!1,valid_types:"StrucWithPotData, NoneType",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"results of the calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"KKRnano output file not found"},{status:302,message:"KKRnano parser retuned an error"}]},class:"aiida_kkr.calculations.kkrnano:KKRnanoCalculation"},"kkr.voro":{description:["AiiDA calculation plugin for a voronoi calculation (creation of starting potential and shapefun)."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Use a node that specifies the input parameters"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_KKR",required:!1,valid_types:"RemoteData, NoneType",info:"Use a node that specifies a parent KKR calculation"},{name:"potential_overwrite",required:!1,valid_types:"SinglefileData, NoneType",info:"Use a node that specifies the potential which is used instead of the voronoi output potential"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"shapefun_overwrite",required:!1,valid_types:"SinglefileData, NoneType",info:"Use a node that specifies the shapefun which is used instead of the voronoi output"},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:"Use a node that specifies the input crystal structure"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"results of the calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"Voronoi output file not found"},{status:302,message:"Voronoi parser retuned an error"}]},class:"aiida_kkr.calculations.voro:VoronoiCalculation"}},"aiida.data":{"kkr.strucwithpot":"aiida_kkr.data.strucwithpot:StrucWithPotData"},"aiida.parsers":{"kkr.kkrimporterparser":"aiida_kkr.parsers.kkrimporter:KkrImporterParser","kkr.kkrimpparser":"aiida_kkr.parsers.kkrimp:KkrimpParser","kkr.kkrnanoparser":"aiida_kkr.parsers.kkrnano:KKRnanoParser","kkr.kkrparser":"aiida_kkr.parsers.kkr:KkrParser","kkr.voroparser":"aiida_kkr.parsers.voro:VoronoiParser"},"aiida.workflows":{"kkr.STM":{description:["Workchain for the Simulation of a (superconducting) STM tip around a magnetic impurity."," The workchain uses the (converged) impurity calculation of a host system and combines them"," with vacuum sites in positions",""," inputs::"," :param options: (Dict), computer options"," :param tip_position: (Dict), specify the position of the STM tip"," :param impurity_info: (Dict), information about the impurity cluster"," :param host_calc: (RemoteData), information about the host structure of the sample"," :param wf_parameters: (Dict), parameters that are used to run the workflow"," :param kkr: (Code), KKR host code for the writing out kkrflex files"," :param kkrimp: (Code), KKR impurity code for the normal state impurity scf and BdG impurity DOS calculation"," :param gf_writeout.params_kkr_overwrite (Dict), overwrite parameters for the GF calculation"," :param kkr_imp_sub.params_kkr_overwrite (Dict), overwrite parameters for the impurity calculation",""," returns::",""," :return workflow_info: (Dict), Information of workflow results"," like success, last result node, list with convergence behavior"," :return STM_dos_data: (XYData), Returns the plot of the lmDOS of the calculation"," :retrun STM_lmdos_data: (XYData), Returns the interpolated lmDOS of the calculation"],spec:{inputs:[{name:"BdG",required:!0,valid_types:"Data",info:""},{name:"gf_writeout",required:!0,valid_types:"Data",info:""},{name:"host_remote",required:!0,valid_types:"RemoteData",info:"Remote Data containing the remote folder from the outputs of the host calculation"},{name:"imp_info",required:!0,valid_types:"Dict",info:"Information of the impurity like position in the unit cell, screening cluster, atom type."},{name:"imp_potential_node",required:!0,valid_types:"SinglefileData",info:"Impurity potential node"},{name:"kkrimp",required:!0,valid_types:"Code",info:"KKRimp code, always needed."},{name:"host_calc",required:!1,valid_types:"RemoteData, NoneType",info:"The information about the clean host structure is required in order to continue the clusterInside a bigger host structure with empty sites."},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` +Initial non-collinear angles for the magnetic moments of the impurities. These values will be written into the \`kkrflex_angle\` input file of KKRimp. +The Dict node should be of the form + initial_noco_angles = Dict({ + 'theta': [theta_at1, theta_at2, ..., theta_atN], # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], # list phi values in degrees (0..360) + 'fix_dir': [True, False, ..., True/False], # list of booleans indicating of the direction of the magentic moment should be fixed or is allowed to be updated (True means keep the direction of the magnetic moment fixed) + }) + +Note: The length of the theta, phi and fix_dir lists have to be equal to the number of atoms in the impurity cluster. +`},{name:"kkr",required:!1,valid_types:"Code, NoneType",info:"KKRhost code, needed if gf_dos_remote is not given."},{name:"kkrflex_files",required:!1,valid_types:"RemoteData, NoneType",info:"with this input we can directly load the gf_dos files without calculating them"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options (resources, quene name, etc.)."},{name:"remote_data",required:!1,valid_types:"RemoteData, NoneType",info:"Remote data from a converged kkr calculation, required for the gf writeout step"},{name:"rimpshift",required:!1,valid_types:"Dict, NoneType",info:` +Shift for atoms in the impurity cluster used in U-transformation. + +The Dict node should be of the form + rimpshift = Dict({'shifts': [[0., 0., 0.], ... ]}) +where the shifts are given in atomic units (i.e. the internal KKR units). + +Note: The length of the 'shifts' attribute should be an array with three numbers indicating the shift for each atom in the impurity cluster. +`},{name:"tip_position",required:!1,valid_types:"Dict, NoneType",info:"How many sites will be scanned in the da and db direction (Bravais Lattice). And the layer that is being scanned."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"Workflow parameter (see `kkr_dos_wc.get_wf_defaults()`)."}],outputs:[{name:"STM_dos_data",required:!0,valid_types:"XyData",info:""},{name:"STM_dos_data_lmdos",required:!0,valid_types:"XyData",info:""},{name:"combined_imp_info",required:!0,valid_types:"Dict",info:""},{name:"combined_imp_potential",required:!0,valid_types:"SinglefileData",info:""},{name:"kkrflexfiles",required:!0,valid_types:"RemoteData",info:""},{name:"tip_position",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The position provided for the STM probe are incorrect"},{status:101,message:"The node provided for the impurity info is not valid"},{status:102,message:"No impurity node has been given in the intput"},{status:103,message:"No impurity info has been given in the input"},{status:104,message:`Neither the kkrflex files nor the KKR builder have been given. +Please provide already converged kkrflex files, or the kkr builder to evaluate them`},{status:201,message:"A step in the kkr_imp_dos workflow has failed"}]},class:"aiida_kkr.workflows.kkr_STM:kkr_STM_wc"},"kkr.bs":{description:["Workchain for BandStructure calculation, starting from RemoteFolderData of the previous converged KKR calculation remote folder data",""," inputs:"," :param wf_parameters: (Dict), (optional); Workchain Specifications, contains nepts, tempr, emin (in eV relative to EF), emax (in eV),"," and RCLUSTZ (can be used to increase the screening cluster radius) keys.",""," :param options: (Dict), (optional); Computer Specifications, scheduler command, parallel or serial"," :param kpoints: (KpointsData),(optional); Kpoints data type from the structure,"," but not mendatory as it can be extracted from structure internaly from the remote data"," :param remote_data: (RemoteData)(mendaory); From the previous kkr-converged calculation."," :param kkr: (Code)(mendaory); KKR code specifiaction"," :param label: (Str) (optional) ; label for WC but will be found in the 'result_wf' output"," Dict as 'BS_wf_label' key"," :param description: (Str) (optional) : description for WC but will be found in the 'result_wf' output"," Dict as 'BS_wf_description' key","",""," returns:"," :out BS_Data : (ArrayData) ; Consist of BlochSpectralFunction, k_points (list), energy_points (list), special_kpoints(dict)"," :out result_wf: (Dict); work_chain_specifications node, BS_data node, remote_folder node"],spec:{inputs:[{name:"kkr",required:!0,valid_types:"Code",info:"KKRhost code, needed to run the qdos KkrCalculation"},{name:"remote_data",required:!0,valid_types:"RemoteData",info:"Parent folder of previously converged KkrCalculation"},{name:"description",required:!1,valid_types:"Str, NoneType",info:"description for the workflow"},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:`Initial non-collinear angles for the magnetic moments. See KkrCalculation for details. + If this is found in the input potentially extracted nonco angles from the parent calulation are overwritten!`},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"K-points data for the calculation. If not given the seekpath library is used to find the irreducable k-points of a structure."},{name:"label",required:!1,valid_types:"Str, NoneType",info:"label for the workflow"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options (walltime etc.) passed onto KkrCalculation"},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Overwrite some input parameters of the parent KKR calculation."},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:` +Settings for running a LDA+U calculation. The Dict node should be of the form + settings_LDAU = Dict(dict={'iatom=0':{ + 'L': 3, # l-block which gets U correction (1: p, 2: d, 3: f-electrons) + 'U': 7., # U value in eV + 'J': 0.75, # J value in eV + 'Eref_EF': 0., # reference energy in eV relative to the Fermi energy. This is the energy where the projector wavefunctions are calculated (should be close in energy where the states that are shifted lie (e.g. for Eu use the Fermi energy)) + }}) + Note: you can add multiple entries like the one for iatom==0 in this example. The atom index refers to the corresponding atom in the impurity cluster. +`},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters of the bandstructure workflow (see output of kkr_bs_wc.get_wf_default() for more details)."}],outputs:[{name:"BS_Data",required:!0,valid_types:"ArrayData",info:""},{name:"results_wf",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:161,message:"No remote_data was provided as Input"},{status:162,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:163,message:"calc_parameters given are not consistent! Hint: did you give an unknown keyword?"},{status:164,message:"calc_parameters not complete"},{status:165,message:"KKR Band Structure calculation failed"},{status:166,message:"No K-POINTS can be extracted from the structure data"},{status:167,message:"No K-POINTS can be extracted from the primtive structure data rather conventional structure data"},{status:168,message:"Input remote_data node neither output of a KKR/voronoi calculation nor of kkr_scf_wc workflow"}]},class:"aiida_kkr.workflows.bs:kkr_bs_wc"},"kkr.combine_imp":{description:["Workchain that combines 2 converged single-impurity calculations to a bigger impurity,"," reusing the preconverged potentials. This is useful, for example, to study co-doping.",""," Inputs:"," :param impurity1_output_node:(Dict), required, output node from singel impurity wc, and should be one of the"," following:"," * for `kkr_imp_wc`: single_imp_worlfow.outputs.workflow_info"," * for `kkr_imp_sub_wc`: single_imp_worlfow.outputs.workflow_info"," * for `KkrimpCalculation`: single_imp_worlfow.outputs.output_parameters",""," :param impurity2_output_node:(Dict), required, output node from second singel impurity wc, and should be one of"," the following:"," * for `kkr_imp_wc`: single_imp_worlfow.outputs.workflow_info"," * for `kkr_imp_sub_wc`: single_imp_worlfow.outputs.workflow_info"," * for `KkrimpCalculation`: single_imp_worlfow.outputs.output_parameters"," :offset_imp2:(Dict), required, offset of the second impurity with respect to the first impurity position."," e.g. {'index:0 or 1}, the replacement by the second impurity will take place at the same cell"," or at the next nearest cell respectively.",""," :param scf.kkrimp: (Code), mandatory, KKRimp code needed to submit kkr_imp_wc"," :param scf.wf_parameters: (Dict), optional, KKRimp code needed to submit kkr_imp_sub_wc"," :param scf.options: (Dict), optional, computer options for kkr_imp_sub_wc",""," :param host_gf.kkr: (Code), optional, KKR code for submit kkr_flex_wc, needed if remote_data_gf is not given"," :param host_gf.options: (Dict), optional, computer options for kkr_flex_wc"," :param host_gf.params_kkr_overwrite: (Dict), optional, needed for kkr calculation for GF writeout",""," :param wf_parameters_overwrite: (Dict), optional, specifications for wf_parameters of kkr_imp_sub_wc as well"," as well as wf_parameters of kkr_flex_wc."," :param gf_host_remote: (RemoteData), optional, remote folder of a previous kkrflex writeout step"," calculations containing the flexfiles and will be used for combine host GF.",""," Returns:"," :return workflow_info: (Dict), Information of workflow results"," :return last_calc_output_parameters: (Dict), link to output parameters of the last called calculation of the"," scf kkr_imp_sub_wc."," :return last_potential: (SingleFileData) link to last output potential of scf kkr_imp_sub_wc step."," :return last_calc_remote: (RemoteData) link to remote data of last called calculation of the scf step."," :return remote_data_gf: (RemoteData) link to GF_host_remote of outputs of kkr_flex_wc e.g. gf_writeou"," step (only present of host GF was generated here)."," :return JijData: (ArrayData) Consists magnetic interaction data among the magnetic impurity atoms,"," such as vector distance(rx, ry, rz) between atoms, spin interaction magnetude J,"," Dzyaloshinskii-Moriya vector magnitude, and Dzyaloshinskii-Moriya vector component(Dx, Dy, Dz)"," :return JijInfo :(Dict) Consists description about the JijData."],spec:{inputs:[{name:"impurity1_output_node",required:!0,valid_types:"Dict",info:"\nOutput node of a single impurity calculation. This can be the output of either the `kkr_imp_wc`, `kkr_imp_sub_wc`\nworkflows or of an `KkrimpCalculation`.\n\nUse these output Dict nodes:\n * for `kkr_imp_wc`: single_imp_workfow.outputs.workflow_info\n * for `kkr_imp_sub_wc`: single_imp_workfow.outputs.workflow_info\n * for `KkrimpCalculation`: single_imp_workfow.outputs.output_parameters\n"},{name:"impurity2_output_node",required:!0,valid_types:"Dict",info:"Output node of second single impurity calculation. See help string of `impurity1_output_node` for more details."},{name:"offset_imp2",required:!0,valid_types:"Dict",info:`Offset of the secon impurity with respect to the first impurity. +Can be given either via the 'vector' or the 'index' keys in the dictionary. +The 'vector' option allows to give the offset vector in cartesian units and +the 'index' option allows to five the offset vector in units of the lattice +vectors of the host system's structure.`},{name:"scf",required:!0,valid_types:"Data",info:""},{name:"gf_host_remote",required:!1,valid_types:"RemoteData, NoneType",info:`RemoteData node of pre-calculated host Green function (i.e. with kkr_flex_wc). +If given then the writeout step of the host GF is omitted.`},{name:"host_gf",required:!1,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"wf_parameters_overwrite",required:!1,valid_types:"Dict, NoneType",info:"To add or edit wf_parameters in scf namespace and add run optioins, if needed"}],outputs:[{name:"last_calc_output_parameters",required:!0,valid_types:"",info:""},{name:"last_calc_remote",required:!0,valid_types:"",info:""},{name:"last_potential",required:!0,valid_types:"",info:""},{name:"workflow_info",required:!0,valid_types:"",info:""},{name:"JijData",required:!1,valid_types:"",info:""},{name:"JijInfo",required:!1,valid_types:"",info:""},{name:"remote_data_gf",required:!1,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:700,message:"The writeout of the host GF failed"},{status:800,message:"Impurity input is not a single impurity calculation."},{status:850,message:"i_neighbor_inplane needs to be positive and bigger than 0 for in-plane neighbors"},{status:900,message:"Host structures of impurity 1 and impurity 2 are not identical!"},{status:950,message:"The impurity calculations have different NSPIN values"},{status:999,message:"ERROR: take a look"}]},class:"aiida_kkr.workflows._combine_imps:combine_imps_wc"},"kkr.decimation":{description:["Workchain a decimation calculation with KKR starting from a thick slab (center potential should be bulk-like).",""," The workchain"," - creates the structure nodes of the substrate and decimation region from thick slab structure"," - creates starting potentials of the slab structure"," - runs auxiliary voronoi steps to get starting setup correctly for the KKR calculations"," - runs the deci-out step in serial"," - runs the decimation step",""," The workflow starts either from a converged thick film of from a previous decimation calculation (skips the structure and starting potential setup and the voronoi steps).",""," The workflow parameters input can be:"," {'nkz' : 30, # number of k-points in z-direction for substrate"," 'nprinc': 4, # number of layer in principle layer"," 'nplayer': 4, # number of principle layers (naez deci: nprinc*nplayer)"," 'dosmode': False, # run DOS calculation"," 'dos_params': {'emin_EF': -5.0, # EMIN-EF in eV"," 'emax_EF': 3.0, # EMAX-EF in eV"," 'nepts': 96, # number of points in contour"," 'tempr': 100, # smearing temperature"," 'kmesh': [50, 50, 50]}, # k-mesh used in dos calculation"," }",""," :param wf_parameters: Dict node with workchain parameters (see kkr_decimation_wc.get_wf_defaults())"," :param options: Dict node with specifications for the computer (used in decimation step only)"," :param remote_data: mandatory RemoteData node of either a parent slab or previous decimation calculation"," :param kkr: mandatory Code node with KKR code for running deci-out and decimation steps"," :param voronoi: Code node that is mandatory if starting from slab calculation. Is the voronoi code for auxiliary calculations"," :param kpoints: KpointsData node that triggers a band structure calculation. The kpoints specify the k-point path along which the bandstructure is computed with the qdos mode of KKRhost."," :param calc_parameters: Dict node that contains KKR parameters which overwrites settings from the slab parent.",""," :returns structure_decimate: StructureData node of the structure of the decimation region."," :returns structure_substrate: StructureData node of the structure of thesubstrate lattice continuation."," :returns out_params_calc_deci_out: Dict node of the output parameters of the deci-out calculation."," :returns out_params_calc_decimate: Dict node of the output parameters of the decimation calculation."," :returns out_remote_calc_decimate: RemoteData node of the decimation calculation."," :returns out_retrieved_calc_decimate: retrieved FolderData node of the decimation calculation."," :returns dos_data: XyData node with the DOS data at finite imaginary part in the energy contour. Only present in DOS mode."," :returns dos_data_interpol: XyData node with the interpolated DOS data onto the real axis. Only present in DOS mode."],spec:{inputs:[{name:"kkr",required:!0,valid_types:"Code",info:"KKRhost code."},{name:"remote_data",required:!0,valid_types:"RemoteData",info:"Parent calculation (either previous decimation calculation, then voronoi steps are skipped or slab calculation)."},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:"If given overwrites KKR parameters starting from slab params (can be used to run DOS for instance)."},{name:"calc_parameters_decimate",required:!1,valid_types:"Dict, NoneType",info:"Overwrite calculation parameters in the decimation step"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"If given this triggers a bandstructure (i.e. qdos) calculation."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options used in the deicmation step (voronoi and deci-out steps run serially but use the walltime given here)."},{name:"options_deci_out",required:!1,valid_types:"Dict, NoneType",info:"Computer options used in the deci-out step (createion of decifile, if not provided, this step will run in serial)."},{name:"shapefun_deci_overwrite",required:!1,valid_types:"SinglefileData, NoneType",info:"Use a node that specifies the shapefun which is used instead of the voronoi output for the decimation step"},{name:"shapefun_substrate_overwrite",required:!1,valid_types:"SinglefileData, NoneType",info:"Use a node that specifies the shapefun which is used instead of the voronoi output for the deci-out step"},{name:"voronoi",required:!1,valid_types:"Code, NoneType",info:"Voronoi code. Only needed if remote_data is slab claculation and not a previous decimation run."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"parameters for decimation setup (used only if not started from previous decimation calculation)."}],outputs:[{name:"out_params_calc_deci_out",required:!0,valid_types:"Dict",info:"Output parameter node of deci-out calculation."},{name:"out_params_calc_decimate",required:!0,valid_types:"Dict",info:"Output parameter node of decimation calculation."},{name:"out_remote_calc_decimate",required:!0,valid_types:"RemoteData",info:"Remote folder of decimation calculation."},{name:"out_retrieved_calc_decimate",required:!0,valid_types:"FolderData",info:"Retrieved folder of decimation calculation."},{name:"structure_decimate",required:!0,valid_types:"StructureData",info:"Structure of decimation region."},{name:"structure_substrate",required:!0,valid_types:"StructureData",info:"Structure of substrate lattice continuation."},{name:"dos_data",required:!1,valid_types:"XyData",info:"DOS data with finite imaginary part in the energy contour."},{name:"dos_data_interpol",required:!1,valid_types:"XyData",info:"interpolated DOS data onto the real axis."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"Given remote_data is not correct (needs to be a slab or decimation KKR calculation)"},{status:301,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:302,message:"The code you provided for voronoi does not use the plugin kkr.voro"},{status:303,message:"The voronoi step for the starting potential of the substrate failed."},{status:304,message:"The voronoi step for the starting potential of the decimation region failed."},{status:305,message:"The deci-out step (writeout of continuation GF) failed."}]},class:"aiida_kkr.workflows._decimation:kkr_decimation_wc"},"kkr.dos":{description:["Workchain a DOS calculation with KKR starting from the remoteData node"," of a previous calculation (either Voronoi or KKR).",""," :param wf_parameters: (Dict); Workchain specifications"," :param options: (Dict); specifications for the computer"," :param remote_data: (RemoteData), mandatory; from a KKR or Vornoi calculation"," :param kkr: (Code), mandatory; KKR code running the dos calculation",""," :return result_kkr_dos_wc: (Dict), Information of workflow results"," like Success, last result node, list with convergence behavior"],spec:{inputs:[{name:"kkr",required:!0,valid_types:"Code",info:"KKRhost Code node used to run the DOS calculation."},{name:"remote_data",required:!0,valid_types:"RemoteData",info:"RemoteData node of the parent calculation."},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:`Initial non-collinear angles for the magnetic moments. See KkrCalculation for details. + If this is found in the input potentially extracted nonco angles from the parent calulation are overwritten!`},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options used by the workflow."},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Overwrite some input parameters of the parent KKR calculation."},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:` +Settings for running a LDA+U calculation. The Dict node should be of the form + settings_LDAU = Dict(dict={'iatom=0':{ + 'L': 3, # l-block which gets U correction (1: p, 2: d, 3: f-electrons) + 'U': 7., # U value in eV + 'J': 0.75, # J value in eV + 'Eref_EF': 0., # reference energy in eV relative to the Fermi energy. This is the energy where the projector wavefunctions are calculated (should be close in energy where the states that are shifted lie (e.g. for Eu use the Fermi energy)) + }}) + Note: you can add multiple entries like the one for iatom==0 in this example. The atom index refers to the corresponding atom in the impurity cluster. +`},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"Workflow parameter (see `kkr_dos_wc.get_wf_defaults()`)."}],outputs:[{name:"results_wf",required:!0,valid_types:"Dict",info:"Results collected by the workflow."},{name:"dos_data",required:!1,valid_types:"XyData",info:"XyData node of the parsed DOS output."},{name:"dos_data_interpol",required:!1,valid_types:"XyData",info:"XyData node of the parsed DOS output, interpolated onto the real axis."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:161,message:"No remote_data was provided as Input"},{status:162,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:163,message:"calc_parameters given are not consistent! Hint: did you give an unknown keyword?"},{status:164,message:"calc_parameters not complete"},{status:165,message:"dos_params given in wf_params are not valid"},{status:166,message:"KKR dos calculation failed"}]},class:"aiida_kkr.workflows.dos:kkr_dos_wc"},"kkr.eos":{description:["Workchain of an equation of states calculation with KKR.",""," Layout of the workflow:"," 1. determine V0, scale_range, etc. from input"," 2. run voro_start for V0 and smallest volume"," 2.1 get minimum for RMTCORE (needs to be fixed for all calculations to be able to compare total energies"," 3. submit kkr_scf calculations for all volumes using RMTCORE setting determined in step 2"," 4. collect results"],spec:{inputs:[{name:"kkr",required:!0,valid_types:"Code",info:` + Code entry for the KKRhost calculations. + `},{name:"structure",required:!0,valid_types:"StructureData",info:` + Initial structure for which the EOS will be calculated. + `},{name:"voronoi",required:!0,valid_types:"Code",info:` + Code entry for the Voronoi calculations. + `},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:` + KKR input parameters. Everything (except structural factors) which + would normally be written in the inputcard. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:` + Workfunction specific parameters, controlling the behavior of the + EOS workchain. + `}],outputs:[{name:"eos_results",required:!0,valid_types:"Dict",info:""},{name:"explicit_kpoints",required:!1,valid_types:"KpointsData",info:""},{name:"get_explicit_kpoints_path_parameters",required:!1,valid_types:"Dict",info:""},{name:"gs_structure",required:!1,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:221,message:"ERROR: inputs invalid"},{status:222,message:"ERROR: need at least 3 successful calculations"},{status:223,message:"ERROR: nsteps is smaller than 3, need at least three data points to do fitting"},{status:224,message:"given fitfunction name not valid"},{status:225,message:"ERROR: kkr_startpot was not successful. Check you inputs."}]},class:"aiida_kkr.workflows.eos:kkr_eos_wc"},"kkr.gf_writeout":{description:["Workchain of a kkr_flex calculation to calculate the Green function with"," KKR starting from the RemoteData node of a previous calculation (either Voronoi or KKR).",""," :param options: (Dict), Workchain specifications"," :param wf_parameters: (Dict), Workflow parameters that deviate from previous KKR RemoteData"," :param remote_data: (RemoteData), mandatory; from a converged KKR calculation"," :param kkr: (Code), mandatory; KKR code running the flexfile writeout"," :param impurity_info: Dict, mandatory: node specifying information"," of the impurities in the system",""," :return workflow_info: (Dict), Information of workflow results"," like success, last result node, list with convergence behavior"," :return GF_host_remote: (RemoteData), host GF of the system"],spec:{inputs:[{name:"impurity_info",required:!0,valid_types:"Dict",info:""},{name:"remote_data",required:!0,valid_types:"RemoteData",info:""},{name:"kkr",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Set some input parameters of the KKR calculation."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"GF_host_remote",required:!0,valid_types:"RemoteData",info:""},{name:"workflow_info",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:101,message:"ERROR: the 'impurity_info' input Dict node could not be used"},{status:102,message:"ERROR: the code you provided for kkr does not use the plugin kkr.kkr"},{status:103,message:"ERROR: No remote_data was provided as Input"},{status:104,message:"ERROR: calc_parameters given are not consistent! Hint: did you give an unknown keyword?"},{status:105,message:"ERROR: calc_parameters misses keys"},{status:106,message:"ERROR: KKR calculation to write out kkrflex files unsuccessful"},{status:107,message:"ERROR: Could not extract value for Fermi level from parent calculation"}]},class:"aiida_kkr.workflows.gf_writeout:kkr_flex_wc"},"kkr.imp":{description:["Workchain of a kkrimp calculation starting either from scratch (with a structure"," and impurity_info node), or with a converged host potential and impurity"," startpotentials, ... to calculate the converged host-impurity potential of the system.",""," :param options: (Dict), Workchain specifications"," :param wf_parameters: (Dict), specifications for the kkr impurity workflow"," :param voro_aux_parameters: (Dict), specification for the auxiliary voronoi calculation for the impurity"," :param kkrimp: (Code), mandatory: KKRimp code converging the host-imp-potential"," :param kkr: (Code), mandatory: KKR code for calculation the host potential"," :param voronoi: (Code), mandatory: Voronoi code to generate the impurity startpot"," :param remote_data_gf: (RemoteData): remote folder of a previous kkrflex"," calculation containing the flexfiles ..."," :param remote_data_host: (RemoteData): remote folder of a converged KKR"," host calculation",""," :return workflow_info: (Dict), Information of workflow results"," :return last_calc_output_parameters: (Dict), output parameters of"," the last called calculation"," :return last_calc_info: (Dict), information of the last called calculation"],spec:{inputs:[{name:"gf_writeout",required:!0,valid_types:"Data",info:""},{name:"impurity_info",required:!0,valid_types:"Dict",info:"Information of the impurity like position in the unit cell, screening cluster, atom type."},{name:"kkrimp",required:!0,valid_types:"Code",info:"KKRimp code used to converge the impurity calculation"},{name:"scf",required:!0,valid_types:"Data",info:""},{name:"voronoi",required:!0,valid_types:"Code",info:"Voronoi code used to create the impurity starting potential."},{name:"kkr",required:!1,valid_types:"Code, NoneType",info:"KKRhost code used to run GF writeout step."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Options for running the codes (walltime etc.)."},{name:"options_voronoi",required:!1,valid_types:"Dict, NoneType",info:"Options for running the Voronoi code (if differing from general `options` node)"},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Set some input parameters of the KKR calculation for the GF writeout step."},{name:"remote_data_gf",required:!1,valid_types:"RemoteData, NoneType",info:"RemoteData node of precomputed host Green function."},{name:"remote_data_gf_Efshift",required:!1,valid_types:"RemoteData, NoneType",info:"RemoteData node of precomputed host Green function with Fermi level shift (overwrite kkrflex_green and tmat files from first remote_data_gf node."},{name:"remote_data_host",required:!1,valid_types:"RemoteData, NoneType",info:"RemoteData node of the converged host calculation. Used to write out the host Green function."},{name:"startpot",required:!1,valid_types:"SinglefileData, NoneType",info:"Set starting potential (e.g. from preconverged calculation"},{name:"voro_aux_parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the auxiliary voronoi starting potential workflow."},{name:"voro_params_overwrite",required:!1,valid_types:"Dict, NoneType",info:"If given, overwrite the some parameters used as input for auxiliary voronoi calculation of starting potential."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the KKRimp selfconsistency workflow."}],outputs:[{name:"last_calc_info",required:!0,valid_types:"Dict",info:""},{name:"last_calc_output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_data_gf",required:!0,valid_types:"RemoteData",info:""},{name:"workflow_info",required:!0,valid_types:"Dict",info:""},{name:"converged_potential",required:!1,valid_types:"SinglefileData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:141,message:"ERROR: one or more of the codes you provided do not use the necessary plugins: kkr.voro, kkr.kkr, kkr.kkrimp"},{status:142,message:"ERROR: since GF writeout step has to be conducted, 'kkrcode' is needed as an input"},{status:143,message:"ERROR: neither converged host remote nor GF writeout remote is given as an input. One of them is needed to proceed with this workflow!"},{status:144,message:"ERROR: sub-workflow for KKRimp convergence failed"},{status:145,message:"ERROR: sub-workflow Kkr_startpot failed (look for failure of voronoi calculation)."}]},class:"aiida_kkr.workflows.kkr_imp:kkr_imp_wc"},"kkr.imp_BdG":{description:["Workchain for one-shot BdG impurity DOS calculation from a converged normal state host calculation."," 1) Normal state impurity potential scf"," 2) One-shot BdG DOS calcluation"," a) Writing out of the BdG kkrflex DOS files"," b) One-shot BdG impurity DOS"," The intermediate steps (1 & 2a) can be skipped by providing the corresponding nodes as inputs to the workflow.",""," inputs::"," :param options: (Dict), computer options"," :param impurity_info: (Dict), information about the impurity cluster"," :param voronoi: (Code), Voronoi code for creating the impurity starting potential"," :param kkr: (Code), KKR host code for the writing out kkrflex files"," :param kkrimp: (Code), KKR impurity code for the normal state impurity scf and BdG impurity DOS calculation"," :param BdG_settings: (Dict), set BdG parameters"," :param imp_scf.startpot: (SinglefileData), pre-converged impurity potential used to start the impurity scf calculation"," :param imp_scf.converged_potential: (SinglefileData), converged impurity potential, skips the impurity scf calculation if provided"," :param imp_scf.wf_parameters: (Dict), parameters for the kkr impurity scf"," :param imp_scf.gf_writeout.params_kkr_overwrite: (Dict), set some input parameters of the KKR calculation for the GF writeout step of impurity scf workchain"," :param imp_scf.gf_writeout.options: (Dict), computer settings"," :param imp_scf.scf.params_overwrite: (Dict), set some input parameters of the KKR impurity scf"," :param imp_scf.options: (Dict), computer settings"," :param imp_scf.remote_data_host: (RemoteData), parent folder of converged host normal state KkrCalculation",""," :param dos.wf_parameters: (Dict), parameters for the DOS calculation"," :param dos.gf_dos_remote: (RemoteData), node of precomputed host GF for DOS energy contour"," :param dos.gf_writeout.params_kkr_overwrite: (Dict), set some input parameters of the KKR calculation for the GF writeout step of imßpurity dos workchain"," :param dos.gf_writeout.host_remote: (RemoteData), parent folder of kkrflex writeout step for DOS calculation"," :param dos.gf_writeout.kkr: (Code), KKR code for writing out of kkrflex files for impurity DOS calculation"," :param dos.gf_writeout.options: (Dict), computer settings"," :param dos.options: (Dict), computer settings",""," returns::"," :return workflow_info: (Dict), Information on workflow results"," :return output_parameters: (Dict), output parameters of the workflow"," :return dos_data: (XyData), impurity DOS data output node"," :return dos_data_interpol: (XyData), interpolated DOS data output node"," :return impurity_potential: (SinglefileData), converged normal state impurity potential node"," :return gf_host_BdG: (RemoteData), kkrflex writeout step files of DOS calculation"],spec:{inputs:[{name:"BdG_scf",required:!0,valid_types:"Data",info:""},{name:"dos",required:!0,valid_types:"Data",info:""},{name:"imp_scf",required:!0,valid_types:"Data",info:""},{name:"kkrimp",required:!0,valid_types:"Code",info:"KKRimp code used to converge the impurity calculation"},{name:"BdG_settings",required:!1,valid_types:"Dict",info:"Define BdG parameters"},{name:"calc_DOS",required:!1,valid_types:"Bool, NoneType",info:"Set this to TRUE to calculate DOS"},{name:"impurity_info",required:!1,valid_types:"Dict, NoneType",info:"Information of the impurity like position in the unit cell, screening cluster, atom type."},{name:"kkr",required:!1,valid_types:"Code, NoneType",info:"KKRhost code, needed to run the KkrCalculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options (walltime etc.) passed onto KkrCalculation, fall back to settings from parent calculation if not given"},{name:"voronoi",required:!1,valid_types:"Code, NoneType",info:"Voronoi code used to create the impurity starting potential."}],outputs:[{name:"impurity_potential",required:!0,valid_types:"SinglefileData",info:""},{name:"dos_data",required:!1,valid_types:"XyData",info:""},{name:"dos_data_interpol",required:!1,valid_types:"XyData",info:""},{name:"dos_data_lm",required:!1,valid_types:"XyData",info:""},{name:"dos_data_lm_interpol",required:!1,valid_types:"XyData",info:""},{name:"gf_host_BdG",required:!1,valid_types:"RemoteData",info:""},{name:"output_parameters",required:!1,valid_types:"Dict",info:""},{name:"workflow_info",required:!1,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:101,message:"The code you provided for kkrimp does not use the plugin kkr.kkrimp"},{status:102,message:"The code you provided for voronoi does not use the plugin kkr.voronoi"},{status:200,message:"Parent calculation is not valid"}]},class:"aiida_kkr.workflows.imp_BdG:kkrimp_BdG_wc"},"kkr.imp_dos":{description:["Workchain of a DOS calculation for an impurity system starting from a"," converged impurity calculation or workflow",""," :param options: (Dict), computer options"," :param wf_parameters: (Dict), specifications for the DOS"," :param kkr: (Code), mandatory: KKR code for gf_writeout step"," :param kkrimp: (Code), mandatory: KKRimp code for DOS calculation"," :param imp_host_pot: (SinglefileData), mandatory: impurity startpotential",""," :return workflow_info: (Dict), Information on workflow results"," :return last_calc_output_parameters: (Dict), output parameters of"," the last called calculation"," :return last_calc_info: (Dict), information of the last called calculation"],spec:{inputs:[{name:"BdG",required:!0,valid_types:"Data",info:""},{name:"gf_writeout",required:!0,valid_types:"Data",info:""},{name:"kkrimp",required:!0,valid_types:"Code",info:"KKRimp code, always needed."},{name:"gf_dos_remote",required:!1,valid_types:"RemoteData, NoneType",info:"RemoteData node of precomputed host GF for DOS energy contour."},{name:"host_remote",required:!1,valid_types:"RemoteData, NoneType",info:"RemoteData node of the (converged) host calculation."},{name:"imp_pot_sfd",required:!1,valid_types:"SinglefileData, NoneType",info:"impurity potential single file data. Needs also impurity_info node."},{name:"impurity_info",required:!1,valid_types:"Dict, NoneType",info:"impurity info node that specifies the relation between imp_pot_sfd to the host system. Mandatory if imp_pot_sfd is given."},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` +Initial non-collinear angles for the magnetic moments of the impurities. These values will be written into the \`kkrflex_angle\` input file of KKRimp. +The Dict node should be of the form + initial_noco_angles = Dict({ + 'theta': [theta_at1, theta_at2, ..., theta_atN], # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], # list phi values in degrees (0..360) + 'fix_dir': [True, False, ..., True/False], # list of booleans indicating of the direction of the magentic moment should be fixed or is allowed to be updated (True means keep the direction of the magnetic moment fixed) + }) + +Note: The length of the theta, phi and fix_dir lists have to be equal to the number of atoms in the impurity cluster. +`},{name:"kkr",required:!1,valid_types:"Code, NoneType",info:"KKRhost code, needed if gf_dos_remote is not given."},{name:"kkrimp_remote",required:!1,valid_types:"RemoteData, NoneType",info:"RemoteData node of previous (converged) KKRimp calculation."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options (resources, quene name, etc.)."},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Set some input parameters of the KKR calculation."},{name:"rimpshift",required:!1,valid_types:"Dict, NoneType",info:` +Shift for atoms in the impurity cluster used in U-transformation. + +The Dict node should be of the form + rimpshift = Dict({'shifts': [[0., 0., 0.], ... ]}) +where the shifts are given in atomic units (i.e. the internal KKR units). + +Note: The length of the 'shifts' attribute should be an array with three numbers indicating the shift for each atom in the impurity cluster. +`},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:"Settings for LDA+U run (see KkrimpCalculation for details)."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"DOS workflow parameters (energy range, etc.)."}],outputs:[{name:"dos_data",required:!0,valid_types:"XyData",info:""},{name:"last_calc_info",required:!0,valid_types:"Dict",info:""},{name:"last_calc_output_parameters",required:!0,valid_types:"Dict",info:""},{name:"workflow_info",required:!0,valid_types:"Dict",info:""},{name:"dos_data_interpol",required:!1,valid_types:"XyData",info:""},{name:"dos_data_interpol_lm",required:!1,valid_types:"XyData",info:""},{name:"dos_data_lm",required:!1,valid_types:"XyData",info:""},{name:"gf_dos_remote",required:!1,valid_types:"RemoteData",info:"RemoteData node of the computed host GF."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:220,message:"Unknown problem detected."},{status:221,message:"Unable to find the parent remote_data node that led to the input impurity calculation. You need to specify `host_remote` and `impurity_info` nodes."},{status:222,message:"The gf_writeout workflow was not succesful, cannot continue."},{status:223,message:"The input nodes `imp_pot_sfd` and `kkrimp_remote` are given but are mutually exclusive"},{status:224,message:"KKRhost code node (`inputs.kkr`) is missing if gf_dos_remote is not given."},{status:225,message:"`host_remote` node is missing if gf_dos_remote is not given."},{status:226,message:"KKRimp sub-workflow failed."}]},class:"aiida_kkr.workflows.kkr_imp_dos:kkr_imp_dos_wc"},"kkr.imp_sub":{description:["Workchain of a kkrimp self consistency calculation starting from the"," host-impurity potential of the system. (Not the entire kkr_imp workflow!)",""," :param options: (Dict), Workchain specifications"," :param wf_parameters: (Dict), specifications for the calculation"," :param host_imp_startpot: (RemoteData), mandatory; input host-impurity potential"," :param kkrimp: (Code), mandatory; KKRimp code converging the host-imp-potential"," :param remote_data: (RemoteData), mandatory; remote folder of a previous"," kkrflex calculation containing the flexfiles ..."," :param kkrimp_remote: (RemoteData), remote folder of a previous kkrimp calculation"," :param impurity_info: (Dict), Parameter node with information"," about the impurity cluster",""," :return workflow_info: (Dict), Information of workflow results"," like success, last result node, list with"," convergence behavior"," :return host_imp_pot: (SinglefileData), output potential of the sytem"],spec:{inputs:[{name:"kkrimp",required:!0,valid_types:"Code",info:""},{name:"host_imp_startpot",required:!1,valid_types:"SinglefileData, NoneType",info:""},{name:"impurity_info",required:!1,valid_types:"Dict, NoneType",info:""},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` +Initial non-collinear angles for the magnetic moments of the impurities. These values will be written into the \`kkrflex_angle\` input file of KKRimp. +The Dict node should be of the form + initial_noco_angles = Dict({ + 'theta': [theta_at1, theta_at2, ..., theta_atN], # list theta values in degrees (0..180) + 'phi': [phi_at1, phi_at2, ..., phi_atN], # list phi values in degrees (0..360) + 'fix_dir': [True, False, ..., True/False], # list of booleans indicating of the direction of the magentic moment should be fixed or is allowed to be updated (True means keep the direction of the magnetic moment fixed) + }) + +Note: The length of the theta, phi and fix_dir lists have to be equal to the number of atoms in the impurity cluster. +`},{name:"kkrimp_remote",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"params_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Dict of parameters that are given to the KKRimpCalculation. Overwrites automatically set values!"},{name:"remote_data",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"remote_data_Efshift",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"rimpshift",required:!1,valid_types:"Dict, NoneType",info:` +Shift for atoms in the impurity cluster used in U-transformation. + +The Dict node should be of the form + rimpshift = Dict({'shifts': [[0., 0., 0.], ... ]}) +where the shifts are given in atomic units (i.e. the internal KKR units). + +Note: The length of the 'shifts' attribute should be an array with three numbers indicating the shift for each atom in the impurity cluster. +`},{name:"settings_LDAU",required:!1,valid_types:"Dict, NoneType",info:"LDA+U settings. See KKRimpCalculation for details."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"workflow_info",required:!0,valid_types:"Dict",info:""},{name:"host_imp_pot",required:!1,valid_types:"SinglefileData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:121,message:"ERROR: Not both host-impurity potential and GF remote found in the inputs. Provide either both of them or a RemoteData from a previous kkrimp calculation."},{status:122,message:"ERROR: The code you provided for KKRimp does not use the plugin kkr.kkrimp"},{status:123,message:"ERROR: Unable to extract parent paremeter node of input remote folder"},{status:124,message:"ERROR: No calculation parameters provided"},{status:125,message:`ERROR: Last KKRcalc in SUBMISSIONFAILED state! +stopping now`},{status:126,message:"ERROR: Maximal number of KKR restarts reached. Exiting now!"},{status:127,message:"ERROR: Last_remote could not be set to a previous succesful calculation"},{status:128,message:"ERROR: There are still missing calculation parameters"},{status:129,message:"ERROR: Parameters could not be updated"},{status:130,message:"ERROR: Last calculation is not in finished state"},{status:131,message:"The input `remote_data` node has no valid calculation parent."},{status:132,message:"The parent calculation of the input `remote_data` node was not succesful."},{status:133,message:"ERROR: Last calculation does not have an output potential."}]},class:"aiida_kkr.workflows.kkr_imp_sub:kkr_imp_sub_wc"},"kkr.jij":{description:["Workchain for calculation of exchange coupling constants Jij and Dij if parent calculation used the SOC solver.",""," inputs::",""," :param wf_parameters: optional Dict node of workchain specifications, contains settings like Jij radius cutoff,"," selection of sites for i and j and numerical cutoffs. None values in the accuracy sub-dict"," means that values from parent calculation are coptied."," :param remote_data: mandatory RemoteData node of parent (i.e. converged) KkrCalculation"," :param kkr: optional Code for KKRhost executable (if not given the same as in the parent calculation is used)"," :param options: optional Dict computer options like scheduler command or parallelization",""," returns::",""," :return jij_data: ArrayData with the arrays 'Jij_expanded' (Table of all Jij and Dij pairs) and 'positions_expanded' (positions of all ij pairs)"," :return structure_jij_sites: StructureData"],spec:{inputs:[{name:"kkr",required:!0,valid_types:"Code",info:"KKRhost code, needed to run the Jij KkrCalculation"},{name:"remote_data",required:!0,valid_types:"RemoteData",info:"Parent folder of previously converged KkrCalculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options (walltime etc.) passed onto KkrCalculation, fall back to settings from parent calculation if not given"},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Overwrite some input parameters of the parent KKR calculation."},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters of the bandstructure workflow (see output of kkr_bs_wc.get_wf_default() for more details)."}],outputs:[{name:"jij_data",required:!0,valid_types:"ArrayData",info:""},{name:"results_wf",required:!0,valid_types:"Dict",info:""},{name:"structure_jij_sites",required:!0,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:160,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:161,message:"Parent calculation is not valid"},{status:162,message:"KKR Band Structure calculation failed"},{status:163,message:"Parsing of Jij calculations failed"}]},class:"aiida_kkr.workflows.jijs:kkr_jij_wc"},"kkr.scf":{description:["Workchain for converging a KKR calculation (SCF).",""," It converges the charge potential."," Two paths are possible:",""," (1) Start from a structure and run a voronoi calculation first,"," optional with calc_parameters"," (2) Start from an existing Voronoi or KKR calculation, with a remoteData",""," :param wf_parameters: (Dict), Workchain Specifications"," :param options: (Dict); specifications for the computer"," :param structure: (StructureData), Crystal structure"," :param calc_parameters: (Dict), Voronoi/Kkr Parameters"," :param remote_data: (RemoteData), from a KKR, or Voronoi calculation"," :param voronoi: (Code)"," :param kkr: (Code)",""," :return output_kkr_scf_wc_para: (Dict), Information of workflow results"," like Success, last result node, list with convergence behavior",""," minimum input example:"," 1. Code1, Code2, Structure, (Parameters), (wf_parameters)"," 2. Code2, remote_data, (Parameters), (wf_parameters)",""," maximum input example:"," 1. Code1, Code2, Structure, Parameters"," wf_parameters: {'queue_name' : String,",` 'resources' : dict({"num_machines": int, "num_mpiprocs_per_machine" : int})`," 'walltime' : int}"," 2. Code2, (remote-data), wf_parameters as in 1.",""," Hints:"," 1. This workflow does not work with local codes!"],spec:{inputs:[{name:"kkr",required:!0,valid_types:"Code",info:"KKRhost code node which will run the KkrCalculations"},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:` + KKR-specific calculation parameters (LMAX etc.), + usually set up with the help of the \`kkrparams\` class. + `},{name:"initial_noco_angles",required:!1,valid_types:"Dict, NoneType",info:` + Initial non-collinear angles for the magnetic moments of the + impurities. See KkrCalculation for details. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:` + Computer settings used by the calculations in the workflow + (see also help string of wf_parameters). + `},{name:"params_kkr_overwrite",required:!1,valid_types:"Dict, NoneType",info:"Set some input parameters of the KKR calculation."},{name:"remote_data",required:!1,valid_types:"RemoteData, NoneType",info:` + RemoteFolder node of a preconverged calculation. + Can be used as a starting point to skip the Voronoi step. + `},{name:"startpot_overwrite",required:!1,valid_types:"SinglefileData, NoneType",info:` + Potential SinglefileData, can be used to overwrite + the starting potential from Voronoi + (the shapefun will be used though and thus needs to be compatible). + This can be used to construct a better starting potential from a + preconverged calculation (e.g. in a smaller unit cell). + `},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:` + Input structure for which a calculation is started with a + VoronoiCalculation. + Can be skipped if a previous KkrCalculation is given with the + \`remote_data\` input node. + `},{name:"voronoi",required:!1,valid_types:"Code, NoneType",info:"\n Voronoi code node, needed only if `structure` input node is given.\n "},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:` + Settings for the workflow. Use \`KkrCalculation.get_wf_defaults()\` + to get the default values and default options. + `}],outputs:[{name:"output_kkr_scf_wc_ParameterResults",required:!0,valid_types:"Dict",info:""},{name:"final_dosdata_interpol",required:!1,valid_types:"XyData",info:""},{name:"last_InputParameters",required:!1,valid_types:"Dict",info:""},{name:"last_RemoteData",required:!1,valid_types:"RemoteData",info:""},{name:"last_calc_out",required:!1,valid_types:"Dict",info:""},{name:"last_noco_angles",required:!1,valid_types:"Dict",info:""},{name:"results_vorostart",required:!1,valid_types:"Dict",info:""},{name:"starting_dosdata_interpol",required:!1,valid_types:"XyData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:221,message:"Unable to extract parent paremeter node of input remote folder"},{status:222,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:223,message:"The code you provided for voronoi does not use the plugin kkr.voro"},{status:224,message:"ERROR: StructureData was provided, but no voronoi code was provided"},{status:225,message:"ERROR: No StructureData nor remote_data was provided as Input"},{status:226,message:"ERROR: kkr_startpot_wc step failed!"},{status:227,message:"DOS run unsuccessful. Check inputs."},{status:228,message:"ERROR: calc_parameters given are not consistent! Missing mandatory keys"},{status:229,message:"ERROR: calc_parameters given are not consistent! Hint: did you give an unknown keyword?"},{status:230,message:"ERROR: calc_parameters not given as input but are needed!"},{status:231,message:"ERROR: parameter update unsuccessful: some key, value pair not valid!"},{status:232,message:"ERROR: calc_parameters misses keys"},{status:233,message:"ERROR: last_remote could not be set to a previous successful calculation"},{status:234,message:"ERROR: maximal number of KKR restarts reached. Exiting now!"},{status:235,message:"ERROR: last KKRcalc in SUBMISSIONFAILED state"}]},class:"aiida_kkr.workflows.kkr_scf:kkr_scf_wc"},"kkr.startpot":{description:["Workchain create starting potential for a KKR calculation by running"," voronoi and getting the starting DOS for first checks on the validity of the input setting."," Starts from a structure together with a KKR parameter node.",""," :param wf_parameters: (Dict), Workchain specifications"," :param options: (Dict), specifications for the computer"," :param structure: (StructureData), aiida structure node to begin"," calculation from (needs to contain vacancies, if KKR needs empty spheres)"," :param kkr: (Code)"," :param voronoi: (Code)"," :param calc_parameters: (Dict), KKR parameter set, passed on to voronoi run.",""," :return result_kkr_startpot_wc: (Dict), Information of workflow results"," like Success, last result node, dos array data"],spec:{inputs:[{name:"voronoi",required:!0,valid_types:"Code",info:"Voronoi code"},{name:"calc_parameters",required:!1,valid_types:"Dict, NoneType",info:` + KKR-specific parameters passed onto the VoronoiCalculation (lmax etc.). + `},{name:"kkr",required:!1,valid_types:"Code, NoneType",info:"Kkr code, only needed only if DOS is calculated."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Computer options passed onto the calculations"},{name:"parent_KKR",required:!1,valid_types:"RemoteData, NoneType",info:` + RemoteData node of a KKR calculation which is used to overwrite + the output potential Voronoi constructs + (typically used to increase lmax). Cannot be used with a different + structure in the input since the structure is extracted from the + parent_KKR. + `},{name:"startpot_overwrite",required:!1,valid_types:"SinglefileData, NoneType",info:` + Potential which can be used instead of the output potential + Voronoi constructs. + `},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:` + Structure for which the starting potential should be constructed, + not needed if parent_KKR is given (typically used to increase the + lmax but use the output potential of the parent_KKR as starting + potential). + `},{name:"wf_parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters that control the behavior of the workflow"}],outputs:[{name:"results_vorostart_wc",required:!0,valid_types:"Dict",info:""},{name:"last_doscal_dosdata",required:!1,valid_types:"XyData",info:""},{name:"last_doscal_dosdata_interpol",required:!1,valid_types:"XyData",info:""},{name:"last_doscal_results",required:!1,valid_types:"Dict",info:""},{name:"last_params_voronoi",required:!1,valid_types:"Dict",info:""},{name:"last_voronoi_remote",required:!1,valid_types:"RemoteData",info:""},{name:"last_voronoi_results",required:!1,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"The code you provided for kkr does not use the plugin kkr.kkr"},{status:202,message:"The code you provided for voronoi does not use the plugin kkr.voro"},{status:203,message:"Voronoi calculation unsuccessful. Check inputs"},{status:204,message:"Voronoi calculation unsuccessful. Check inputs."},{status:205,message:"Voronoi calculation unsuccessful. Structure inconsistent. Maybe you need empty spheres?"},{status:206,message:"DOS run unsuccessful. Check inputs."},{status:207,message:"Can only take either structure or parent_KKR as input."},{status:208,message:"Need either structure or parent_KKR as input."}]},class:"aiida_kkr.workflows.voro_start:kkr_startpot_wc"}},console_scripts:{"aiida-kkr":"aiida_kkr.cmdline:cmd_root"}},commits_count:87,errors:[],warnings:["W005: Development status in classifiers (beta) does not match development_status in metadata (stable)","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:5},{colorclass:"brown",text:"Parsers",count:5},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:14},{colorclass:"purple",text:"Console scripts",count:1}],pip_install_cmd:"pip install aiida-kkr",is_installable:"True"},"aiida-lammps":{code_home:"https://github.com/aiidaplugins/aiida-lammps",development_status:"beta",entry_point_prefix:"lammps",pip_url:"git+https://github.com/aiidaplugins/aiida-lammps",name:"aiida-lammps",package_name:"aiida_lammps",hosted_on:"github.com",metadata:{classifiers:["Development Status :: 4 - Beta","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3","Programming Language :: Python :: 3.9","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics","Framework :: AiiDA"],author:"Chris Sewell",author_email:"chrisj_sewell@hotmail.com"},aiida_version:"~=2.3",entry_points:{"aiida.calculations":{"lammps.base":{description:["A basic plugin for performing calculations in ``LAMMPS`` using aiida.",""," The plugin will take the input parameters validate them against a schema"," and then use them to generate the ``LAMMPS`` input file. The input file"," is generated depending on the parameters provided, the type of potential,"," the input structure and whether or not a restart file is provided."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Parameters that control the input script generated for the ``LAMMPS`` calculation"},{name:"potential",required:!0,valid_types:"LammpsPotentialData",info:"Potential used in the ``LAMMPS`` calculation"},{name:"structure",required:!0,valid_types:"StructureData",info:"Structure used in the ``LAMMPS`` calculation"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"input_restartfile",required:!1,valid_types:"SinglefileData, NoneType",info:"Input restartfile to continue from a previous ``LAMMPS`` calculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"An optional working directory of a previously completed calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional settings that control the ``LAMMPS`` calculation"}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results",required:!0,valid_types:"Dict",info:"The data extracted from the lammps output file"},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"time_dependent_computes",required:!0,valid_types:"ArrayData",info:"The data with the time dependent computes parsed from the lammps.out"},{name:"trajectories",required:!0,valid_types:"LammpsTrajectory",info:"The data extracted from the lammps trajectory file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"restartfile",required:!1,valid_types:"SinglefileData",info:"The restartfile of a ``LAMMPS`` calculation"},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"the retrieved folder data node could not be accessed."},{status:302,message:"the stdout output file was not found"},{status:303,message:"the stderr output file was not found"},{status:304,message:"the output file is missing, it is possible that LAMMPS never ran"},{status:305,message:"the file with the lammps log was not found"},{status:306,message:"the file with the final variables was not found"},{status:307,message:"the file with the trajectories was not found"},{status:308,message:"the file with the restart information was not found"},{status:309,message:"The parser detected the lammps error :{error}"},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:401,message:"The energy tolerance was not reached at minimization."},{status:402,message:"The force tolerance was not reached at minimization."},{status:1001,message:"error parsing the output file has failed."},{status:1002,message:"error parsing the final variable file has failed."}]},class:"aiida_lammps.calculations.base:LammpsBaseCalculation"},"lammps.raw":{description:["Plugin with minimal interface to run LAMMPS."],spec:{inputs:[{name:"script",required:!0,valid_types:"SinglefileData",info:"Complete input script to use. If specified, `structure`, `potential` and `parameters` are ignored."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"filenames",required:!1,valid_types:"Dict, NoneType",info:"Optional namespace to specify with which filenames the files of ``files`` input should be written."},{name:"files",required:!1,valid_types:"SinglefileData",info:"Optional files that should be written to the working directory."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional settings that control the ``LAMMPS`` calculation"}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results",required:!0,valid_types:"Dict",info:"The data extracted from the lammps out file"},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:309,message:"The parser detected the lammps error :{error}"},{status:351,message:"the file with the lammps out was not found"},{status:1001,message:"parsing the output file has failed."}]},class:"aiida_lammps.calculations.raw:LammpsRawCalculation"}},"aiida.parsers":{"lammps.base":"aiida_lammps.parsers.base:LammpsBaseParser","lammps.raw":"aiida_lammps.parsers.raw:LammpsRawParser"},"aiida.data":{"lammps.potential":"aiida_lammps.data.potential:LammpsPotentialData","lammps.trajectory":"aiida_lammps.data.trajectory:LammpsTrajectory"},"aiida.workflows":{"lammps.base":{description:["Base workchain for calculations using LAMMPS"],spec:{inputs:[{name:"lammps",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"store_restart",required:!1,valid_types:"Bool, NoneType",info:` + Whether to store the restartfile in the repository. + `}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results",required:!0,valid_types:"Dict",info:"The data extracted from the lammps output file"},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"time_dependent_computes",required:!0,valid_types:"ArrayData",info:"The data with the time dependent computes parsed from the lammps.out"},{name:"trajectories",required:!0,valid_types:"LammpsTrajectory",info:"The data extracted from the lammps trajectory file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"restartfile",required:!1,valid_types:"SinglefileData",info:"The restartfile of a ``LAMMPS`` calculation"},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:` + The calculation failed with an unidentified unrecoverable error. + `},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_lammps.workflows.base:LammpsBaseWorkChain"},"lammps.relax":{description:["Workchain to relax a structure using the LAMMPS minimization procedure."],spec:{inputs:[{name:"lammps",required:!0,valid_types:"Data",info:""},{name:"relax",required:!0,valid_types:"",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"store_restart",required:!1,valid_types:"Bool, NoneType",info:` + Whether to store the restartfile in the repository. + `}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results",required:!0,valid_types:"Dict",info:"The data extracted from the lammps output file"},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"time_dependent_computes",required:!0,valid_types:"ArrayData",info:"The data with the time dependent computes parsed from the lammps.out"},{name:"trajectories",required:!0,valid_types:"LammpsTrajectory",info:"The data extracted from the lammps trajectory file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"restartfile",required:!1,valid_types:"SinglefileData",info:"The restartfile of a ``LAMMPS`` calculation"},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:403,message:"The underlying LammpsBaseWorkChain failed"}]},class:"aiida_lammps.workflows.relax:LammpsRelaxWorkChain"},"lammps.md":{description:["Workchain to perform a LAMMPS MD simulation."],spec:{inputs:[{name:"lammps",required:!0,valid_types:"Data",info:""},{name:"md",required:!0,valid_types:"",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"store_restart",required:!1,valid_types:"Bool, NoneType",info:` + Whether to store the restartfile in the repository. + `}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results",required:!0,valid_types:"Dict",info:"The data extracted from the lammps output file"},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"time_dependent_computes",required:!0,valid_types:"ArrayData",info:"The data with the time dependent computes parsed from the lammps.out"},{name:"trajectories",required:!0,valid_types:"LammpsTrajectory",info:"The data extracted from the lammps trajectory file"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"restartfile",required:!1,valid_types:"SinglefileData",info:"The restartfile of a ``LAMMPS`` calculation"},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:403,message:"The underlying LammpsBaseWorkChain failed"}]},class:"aiida_lammps.workflows.molecular_dynamics:LammpsMDWorkChain"}}},commits_count:60,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:2},{colorclass:"green",text:"Workflows",count:3}],pip_install_cmd:"pip install git+https://github.com/aiidaplugins/aiida-lammps",is_installable:"True"},"aiida-lsmo":{code_home:"https://github.com/lsmo-epfl/aiida-lsmo",development_status:"stable",entry_point_prefix:"lsmo",pip_url:"git+https://github.com/lsmo-epfl/aiida-lsmo",name:"aiida-lsmo",package_name:"aiida_lsmo",hosted_on:"github.com",metadata:{author:"Aliaksandr Yakutovich, Daniele Ongari, Leopold Talirz",author_email:"aliaksandr.yakutovich@epfl.ch",version:"1.0.0",description:"AiiDA workflows for the LSMO laboratory at EPFL",classifiers:["Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7"]},aiida_version:">=1.0.0",entry_points:{"aiida.calculations":{"lsmo.ff_builder":"aiida_lsmo.calcfunctions:ff_builder","lsmo.calc_ch4_working_cap":"aiida_lsmo.calcfunctions:calc_ch4_working_cap","lsmo.calc_h2_working_cap":"aiida_lsmo.calcfunctions:calc_h2_working_cap","lsmo.calc_o2_working_cap":"aiida_lsmo.calcfunctions:calc_o2_working_cap","lsmo.calc_selectivity":"aiida_lsmo.calcfunctions:calc_selectivity"},"aiida.parsers":{"lsmo.cp2k_bsse_parser":"aiida_lsmo.parsers:Cp2kBsseParser","lsmo.cp2k_advanced_parser":"aiida_lsmo.parsers:Cp2kAdvancedParser"},"aiida.workflows":{"lsmo.binding_site":"aiida_lsmo.workchains:BindingSiteWorkChain","lsmo.cp2k_binding_energy":"aiida_lsmo.workchains.cp2k_binding_energy:Cp2kBindingEnergyWorkChain","lsmo.cp2k_multistage":"aiida_lsmo.workchains:Cp2kMultistageWorkChain","lsmo.cp2k_multistage_ddec":"aiida_lsmo.workchains:Cp2kMultistageDdecWorkChain","lsmo.isotherm":"aiida_lsmo.workchains:IsothermWorkChain","lsmo.isotherm_multi_temp":"aiida_lsmo.workchains:IsothermMultiTempWorkChain","lsmo.isotherm_calc_pe":"aiida_lsmo.workchains:IsothermCalcPEWorkChain","lsmo.zeopp_multistage_ddec":"aiida_lsmo.workchains:ZeoppMultistageDdecWorkChain","lsmo.sim_annealing":"aiida_lsmo.workchains.sim_annealing:SimAnnealingWorkChain","lsmo.nanoporous_screening_1":"aiida_lsmo.workchains:NanoporousScreening1WorkChain"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-lsmo
Collecting git+https://github.com/lsmo-epfl/aiida-lsmo
+  Cloning https://github.com/lsmo-epfl/aiida-lsmo to /tmp/pip-req-build-m59d81jw
+  Running command git clone --filter=blob:none --quiet https://github.com/lsmo-epfl/aiida-lsmo /tmp/pip-req-build-m59d81jw
+  Resolved https://github.com/lsmo-epfl/aiida-lsmo to commit 6bf08fa42e545dadf889ea8095d7fcdd8d1be15c
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-lsmo to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-lsmo==1.0.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-lsmo 1.0.0 depends on aiida-core~=1.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:5},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"green",text:"Workflows",count:10}],pip_install_cmd:"pip install git+https://github.com/lsmo-epfl/aiida-lsmo",is_installable:"False"},"aiida-metavo-scheduler":{code_home:"https://github.com/pzarabadip/aiida-metavo-scheduler",development_status:"stable",entry_point_prefix:"metavo_scheduler",pip_url:"git+https://github.com/pzarabadip/aiida-metavo-scheduler",name:"aiida-metavo-scheduler",package_name:"aiida_metavo_scheduler",hosted_on:"github.com",metadata:{author:"Pezhman Zarabadi-Poor",author_email:"pzarabadip@gmail.com",version:"1.0.0",description:"",classifiers:["Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Scientific/Engineering"]},aiida_version:">=1.0.0,<1.6",entry_points:{"aiida.cmdline.computer.configure":{sshmetavo:"aiida_metavo_scheduler.metavo.ssh_metavo:CONFIGURE_SSH_CMD"},"aiida.schedulers":{pbsprometavo:"aiida_metavo_scheduler.metavo.pbspro_metavo:PbsproSchedulerMetaVO"},"aiida.transports":{sshmetavo:"aiida_metavo_scheduler.metavo.ssh_metavo:SshTransport"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-metavo-scheduler
Collecting git+https://github.com/pzarabadip/aiida-metavo-scheduler
+  Cloning https://github.com/pzarabadip/aiida-metavo-scheduler to /tmp/pip-req-build-g6epo121
+  Running command git clone --filter=blob:none --quiet https://github.com/pzarabadip/aiida-metavo-scheduler /tmp/pip-req-build-g6epo121
+  Resolved https://github.com/pzarabadip/aiida-metavo-scheduler to commit 955697497641ca13e997431a8e925df6ec3a9eea
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-metavo-scheduler to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-metavo-scheduler==1.0.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-metavo-scheduler 1.0.0 depends on aiida_core<1.6 and >=1.0.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'sshmetavo' does not start with prefix 'metavo_scheduler.'","W010: Entry point 'pbsprometavo' does not start with prefix 'metavo_scheduler.'","W010: Entry point 'sshmetavo' does not start with prefix 'metavo_scheduler.'"],summaryinfo:[{colorclass:"orange",text:"Other (Cmdline computer configure, Schedulers, Transports)",count:3}],pip_install_cmd:"pip install git+https://github.com/pzarabadip/aiida-metavo-scheduler",is_installable:"False"},"aiida-mlip":{entry_point_prefix:"mlip",plugin_info:"https://raw.githubusercontent.com/stfc/aiida-mlip/main/pyproject.toml",code_home:"https://github.com/stfc/aiida-mlip",documentation_url:"https://stfc.github.io/aiida-mlip/",pip_url:"aiida-mlip",name:"aiida-mlip",package_name:"aiida_mlip",hosted_on:"github.com",metadata:{release_date:"2024-06-10",description:"machine learning interatomic potentials aiida plugin",author:"Federica Zanca",author_email:"federica.zanca@stfc.ac.uk",home_page:"https://github.com/stfc/aiida-mlip/",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: BSD License","Natural Language :: English","Programming Language :: Python","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9"],version:"0.2.1"},aiida_version:">=2.5,<3.0",entry_points:{"aiida.calculations":{"mlip.md":{description:["Calcjob implementation to run geometry MD calculations using mlips.",""," Methods"," -------"," define(spec: CalcJobProcessSpec) -> None:"," Define the process specification, its inputs, outputs and exit codes."," prepare_for_submission(folder: Folder) -> CalcInfo:"," Create the input files for the `CalcJob`."],spec:{inputs:[{name:"arch",required:!1,valid_types:"Str, NoneType",info:"Mlip architecture to use for calculation, defaults to mace"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"config",required:!1,valid_types:"JanusConfigfile, NoneType",info:"Name of the log output file"},{name:"device",required:!1,valid_types:"Str, NoneType",info:"Device on which to run calculation (cpu, cuda or mps)"},{name:"ensemble",required:!1,valid_types:"Str, NoneType",info:"Name for thermodynamic ensemble"},{name:"log_filename",required:!1,valid_types:"Str, NoneType",info:"Name of the log output file"},{name:"md_kwargs",required:!1,valid_types:"Dict, NoneType",info:"Keywords for molecular dynamics"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"model",required:!1,valid_types:"ModelData, NoneType",info:"Mlip model used for calculation"},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"precision",required:!1,valid_types:"Str, NoneType",info:"Precision level for calculation"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"struct",required:!1,valid_types:"StructureData, NoneType",info:"The input structure."}],outputs:[{name:"final_structure",required:!0,valid_types:"StructureData",info:""},{name:"log_output",required:!0,valid_types:"SinglefileData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results_dict",required:!0,valid_types:"Dict",info:"The `results_dict` output node of the successful calculation."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stats_file",required:!0,valid_types:"SinglefileData",info:""},{name:"std_output",required:!0,valid_types:"SinglefileData",info:""},{name:"summary",required:!0,valid_types:"SinglefileData",info:""},{name:"traj_file",required:!0,valid_types:"SinglefileData",info:""},{name:"traj_output",required:!0,valid_types:"TrajectoryData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:305,message:"Some output files missing or cannot be read"}]},class:"aiida_mlip.calculations.md:MD"},"mlip.opt":{description:["Calcjob implementation to run geometry optimisation calculations using mlips.",""," Methods"," -------"," define(spec: CalcJobProcessSpec) -> None:"," Define the process specification, its inputs, outputs and exit codes."," prepare_for_submission(folder: Folder) -> CalcInfo:"," Create the input files for the `CalcJob`."],spec:{inputs:[{name:"arch",required:!1,valid_types:"Str, NoneType",info:"Mlip architecture to use for calculation, defaults to mace"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"config",required:!1,valid_types:"JanusConfigfile, NoneType",info:"Name of the log output file"},{name:"device",required:!1,valid_types:"Str, NoneType",info:"Device on which to run calculation (cpu, cuda or mps)"},{name:"fmax",required:!1,valid_types:"Float, NoneType",info:"Maximum force for convergence"},{name:"fully_opt",required:!1,valid_types:"Bool, NoneType",info:"Fully optimise the cell vectors, angles, and atomic positions"},{name:"log_filename",required:!1,valid_types:"Str, NoneType",info:"Name of the log output file"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"model",required:!1,valid_types:"ModelData, NoneType",info:"Mlip model used for calculation"},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"opt_kwargs",required:!1,valid_types:"Dict, NoneType",info:"Other optimisation keywords"},{name:"out",required:!1,valid_types:"Str, NoneType",info:"Name of the xyz output file"},{name:"precision",required:!1,valid_types:"Str, NoneType",info:"Precision level for calculation"},{name:"properties",required:!1,valid_types:"Str, NoneType",info:"Properties to calculate"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"steps",required:!1,valid_types:"Int, NoneType",info:"Number of optimisation steps"},{name:"struct",required:!1,valid_types:"StructureData, NoneType",info:"The input structure."},{name:"traj",required:!1,valid_types:"Str, NoneType",info:"Path to save optimisation frames to"},{name:"vectors_only",required:!1,valid_types:"Bool, NoneType",info:"Optimise cell vectors, as well as atomic positions"}],outputs:[{name:"final_structure",required:!0,valid_types:"StructureData",info:""},{name:"log_output",required:!0,valid_types:"SinglefileData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results_dict",required:!0,valid_types:"Dict",info:"The `results_dict` output node of the successful calculation."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"std_output",required:!0,valid_types:"SinglefileData",info:""},{name:"traj_file",required:!0,valid_types:"SinglefileData",info:""},{name:"traj_output",required:!0,valid_types:"TrajectoryData",info:""},{name:"xyz_output",required:!0,valid_types:"SinglefileData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:305,message:"Some output files missing or cannot be read"}]},class:"aiida_mlip.calculations.geomopt:GeomOpt"},"mlip.sp":{description:["Calcjob implementation to run single point calculations using mlips.",""," Attributes"," ----------"," XYZ_OUTPUT : str"," Default xyz output file name.",""," Methods"," -------"," define(spec: CalcJobProcessSpec) -> None:"," Define the process specification, its inputs, outputs and exit codes."," validate_inputs(value: dict, port_namespace: PortNamespace) -> Optional[str]:"," Check if the inputs are valid."," prepare_for_submission(folder: Folder) -> CalcInfo:"," Create the input files for the `CalcJob`."],spec:{inputs:[{name:"arch",required:!1,valid_types:"Str, NoneType",info:"Mlip architecture to use for calculation, defaults to mace"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"config",required:!1,valid_types:"JanusConfigfile, NoneType",info:"Name of the log output file"},{name:"device",required:!1,valid_types:"Str, NoneType",info:"Device on which to run calculation (cpu, cuda or mps)"},{name:"log_filename",required:!1,valid_types:"Str, NoneType",info:"Name of the log output file"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"model",required:!1,valid_types:"ModelData, NoneType",info:"Mlip model used for calculation"},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"out",required:!1,valid_types:"Str, NoneType",info:"Name of the xyz output file"},{name:"precision",required:!1,valid_types:"Str, NoneType",info:"Precision level for calculation"},{name:"properties",required:!1,valid_types:"Str, NoneType",info:"Properties to calculate"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"struct",required:!1,valid_types:"StructureData, NoneType",info:"The input structure."}],outputs:[{name:"log_output",required:!0,valid_types:"SinglefileData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results_dict",required:!0,valid_types:"Dict",info:"The `results_dict` output node of the successful calculation."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"std_output",required:!0,valid_types:"SinglefileData",info:""},{name:"xyz_output",required:!0,valid_types:"SinglefileData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:305,message:"Some output files missing or cannot be read"}]},class:"aiida_mlip.calculations.singlepoint:Singlepoint"},"mlip.train":{description:["Calcjob implementation to train mlips.",""," Attributes"," ----------"," DEFAULT_OUTPUT_FILE : str"," Default stdout file name.",""," Methods"," -------"," define(spec: CalcJobProcessSpec) -> None:"," Define the process specification, its inputs, outputs and exit codes."," validate_inputs(value: dict, port_namespace: PortNamespace) -> Optional[str]:"," Check if the inputs are valid."," prepare_for_submission(folder: Folder) -> CalcInfo:"," Create the input files for the `CalcJob`."],spec:{inputs:[{name:"mlip_config",required:!0,valid_types:"JanusConfigfile",info:"Config file with parameters for training"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"fine_tune",required:!1,valid_types:"Bool, NoneType",info:"Whether fine-tuning a model"},{name:"foundation_model",required:!1,valid_types:"ModelData, NoneType",info:"Model to fine-tune"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"checkpoints",required:!0,valid_types:"FolderData",info:""},{name:"compiled_model",required:!0,valid_types:"SinglefileData",info:""},{name:"logs",required:!0,valid_types:"FolderData",info:""},{name:"model",required:!0,valid_types:"ModelData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"results_dict",required:!0,valid_types:"Dict",info:"The `results_dict` output node of the training."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:305,message:"Some output files missing or cannot be read"}]},class:"aiida_mlip.calculations.train:Train"}},"aiida.data":{"mlip.config":"aiida_mlip.data.config:JanusConfigfile","mlip.modeldata":"aiida_mlip.data.model:ModelData"},"aiida.parsers":{"mlip.md_parser":"aiida_mlip.parsers.md_parser:MDParser","mlip.opt_parser":"aiida_mlip.parsers.opt_parser:GeomOptParser","mlip.sp_parser":"aiida_mlip.parsers.sp_parser:SPParser","mlip.train_parser":"aiida_mlip.parsers.train_parser:TrainParser"}},commits_count:70,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:4},{colorclass:"brown",text:"Parsers",count:4},{colorclass:"red",text:"Data",count:2}],pip_install_cmd:"pip install aiida-mlip",is_installable:"True"},"aiida-mpds":{code_home:"https://github.com/mpds-io/mpds-aiida",development_status:"beta",documentation_url:"https://github.com/mpds-io/mpds-aiida",entry_point_prefix:"mpds",pip_url:"git+https://github.com/mpds-io/mpds-aiida",name:"aiida-mpds",package_name:"aiida_mpds",hosted_on:"github.com",metadata:{author:"Andrey Sobolev",author_email:"as@tilde.pro",version:"",description:"Aiida workflows for MPDS based on CRYSTAL",classifiers:["Programming Language :: Python","Programming Language :: Python :: 3.5","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics","Topic :: Scientific/Engineering :: Information Analysis","Framework :: AiiDA"]},aiida_version:">=1.0.1",entry_points:{"aiida.workflows":{"crystal.mpds":"mpds_aiida.workflows.mpds:MPDSStructureWorkchain","crystal.cif":"mpds_aiida.workflows.cif:CIFStructureWorkchain","crystal.aiida":"mpds_aiida.workflows.aiida:AiidaStructureWorkchain"}},commits_count:4,errors:[`E001: Failed to install plugin aiida-mpds
Collecting git+https://github.com/mpds-io/mpds-aiida
+  Cloning https://github.com/mpds-io/mpds-aiida to /tmp/pip-req-build-sw4od4h5
+  Running command git clone --filter=blob:none --quiet https://github.com/mpds-io/mpds-aiida /tmp/pip-req-build-sw4od4h5
+  Resolved https://github.com/mpds-io/mpds-aiida to commit b4ab96db1472969e89159b2b2d76a9836b99c94a
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs (from mpds-aiida==0.10.1)
+  Cloning https://github.com/mpds-io/mpds-ml-labs to /tmp/pip-install-2y_nm2nh/mpds-ml-labs_601d7a0c154742ebb1990bde07b054db
+  Running command git clone --filter=blob:none --quiet https://github.com/mpds-io/mpds-ml-labs /tmp/pip-install-2y_nm2nh/mpds-ml-labs_601d7a0c154742ebb1990bde07b054db
+  Resolved https://github.com/mpds-io/mpds-ml-labs to commit 46b15d3fb27a1ca3389d9fbe44b9061740291bc7
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft (from mpds-aiida==0.10.1)
+  Cloning https://github.com/tilde-lab/aiida-crystal-dft to /tmp/pip-install-2y_nm2nh/aiida-crystal-dft_e25f162197a5436aa2c2bb22378b3727
+  Running command git clone --filter=blob:none --quiet https://github.com/tilde-lab/aiida-crystal-dft /tmp/pip-install-2y_nm2nh/aiida-crystal-dft_e25f162197a5436aa2c2bb22378b3727
+  Resolved https://github.com/tilde-lab/aiida-crystal-dft to commit dfa54170debe4be80c9fc8983e826af46afe5916
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+Collecting reentry>=1.3.1 (from mpds-aiida==0.10.1)
+  Using cached reentry-1.3.3-py3-none-any.whl.metadata (10 kB)
+Requirement already satisfied: six>=1.16.0 in /opt/conda/lib/python3.10/site-packages (from mpds-aiida==0.10.1) (1.16.0)
+Requirement already satisfied: numpy>=1.17.5 in /opt/conda/lib/python3.10/site-packages (from mpds-aiida==0.10.1) (1.26.4)
+Collecting ase>=3.19 (from mpds-aiida==0.10.1)
+  Downloading ase-3.23.0-py3-none-any.whl.metadata (3.8 kB)
+Collecting yascheduler>=1.0.12 (from mpds-aiida==0.10.1)
+  Downloading yascheduler-1.2.0-py3-none-any.whl.metadata (12 kB)
+Collecting scipy>=1.6.0 (from ase>=3.19->mpds-aiida==0.10.1)
+  Downloading scipy-1.14.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (60 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 60.8/60.8 kB 4.3 MB/s eta 0:00:00
+Collecting matplotlib>=3.3.4 (from ase>=3.19->mpds-aiida==0.10.1)
+  Downloading matplotlib-3.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB)
+Requirement already satisfied: setuptools>=36.2 in /opt/conda/lib/python3.10/site-packages (from reentry>=1.3.1->mpds-aiida==0.10.1) (70.0.0)
+Requirement already satisfied: click in /opt/conda/lib/python3.10/site-packages (from reentry>=1.3.1->mpds-aiida==0.10.1) (8.1.7)
+Collecting aiohttp~=3.8 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading aiohttp-3.11.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)
+Collecting asyncssh~=2.11 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading asyncssh-2.18.0-py3-none-any.whl.metadata (9.9 kB)
+Collecting asyncstdlib~=3.10 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading asyncstdlib-3.13.0-py3-none-any.whl.metadata (3.7 kB)
+Collecting attrs~=21.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading attrs-21.4.0-py2.py3-none-any.whl.metadata (9.8 kB)
+Collecting azure-identity~=1.10.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading azure_identity-1.10.0-py3-none-any.whl.metadata (54 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.5/54.5 kB 12.9 MB/s eta 0:00:00
+Collecting azure-mgmt-compute~=27.2.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading azure_mgmt_compute-27.2.0-py3-none-any.whl.metadata (53 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 53.1/53.1 kB 12.5 MB/s eta 0:00:00
+Collecting azure-mgmt-network~=20.0.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading azure_mgmt_network-20.0.0-py3-none-any.whl.metadata (63 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 63.9/63.9 kB 21.3 MB/s eta 0:00:00
+Collecting backoff~=2.1.2 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading backoff-2.1.2-py3-none-any.whl.metadata (13 kB)
+Collecting hcloud~=1.17 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading hcloud-1.35.0-py3-none-any.whl.metadata (4.9 kB)
+Collecting pg8000~=1.19 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading pg8000-1.31.2-py3-none-any.whl.metadata (74 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 74.7/74.7 kB 18.0 MB/s eta 0:00:00
+Collecting python-daemon~=2.3 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading python_daemon-2.3.2-py3-none-any.whl.metadata (2.0 kB)
+Requirement already satisfied: typing-extensions>=4.2.0 in /opt/conda/lib/python3.10/site-packages (from yascheduler>=1.0.12->mpds-aiida==0.10.1) (4.12.1)
+Collecting upcloud_api~=2.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading upcloud_api-2.6.0-py3-none-any.whl.metadata (7.8 kB)
+Requirement already satisfied: aiida-core>=2.0.2 in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.5.2)
+Collecting pycrystal>=1.0.10 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading pycrystal-1.0.16.tar.gz (28 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting mpds_client>=0.24 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading mpds_client-0.24.tar.gz (9.7 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting pyparsing>2.3.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading pyparsing-3.2.0-py3-none-any.whl.metadata (5.0 kB)
+Collecting spglib==2.0.2 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading spglib-2.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.3 kB)
+Requirement already satisfied: jinja2>=2.10 in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.1.4)
+Collecting jsonschema>=3.0.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading jsonschema-4.23.0-py3-none-any.whl.metadata (7.9 kB)
+Requirement already satisfied: packaging in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (24.0)
+Collecting pycodcif (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading pycodcif-3.0.1.tar.gz (79 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 79.6/79.6 kB 21.5 MB/s eta 0:00:00
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Collecting scikit-learn (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading scikit_learn-1.5.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (13 kB)
+Collecting imblearn (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading imblearn-0.0-py2.py3-none-any.whl.metadata (355 bytes)
+Collecting progressbar (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading progressbar-2.5.tar.gz (10 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+Requirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.13.1)
+Requirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.4.2)
+Requirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.8.1)
+Requirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.18.0)
+Requirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.1.10)
+Requirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.1.0)
+Requirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.16)
+Requirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.20.3)
+Requirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (8.25.0)
+Requirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.18.2)
+Requirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.8)
+Requirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.11.0)
+Requirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.12.0)
+Requirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.21.10)
+Requirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.4)
+Requirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.9.8)
+Requirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.9.9)
+Requirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.7.3)
+Requirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2021.3)
+Requirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.0.1)
+Requirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.32.3)
+Requirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.0.30)
+Requirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.8.10)
+Requirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.66.4)
+Requirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.9.5)
+Requirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.16.0)
+Collecting aiohappyeyeballs>=2.3.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading aiohappyeyeballs-2.4.3-py3-none-any.whl.metadata (6.1 kB)
+Collecting aiosignal>=1.1.2 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading aiosignal-1.3.1-py3-none-any.whl.metadata (4.0 kB)
+Collecting async-timeout<6.0,>=4.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading async_timeout-5.0.1-py3-none-any.whl.metadata (5.1 kB)
+Collecting frozenlist>=1.1.1 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (13 kB)
+Requirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/lib/python3.10/site-packages (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1) (6.0.5)
+Collecting propcache>=0.2.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading propcache-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)
+Collecting yarl<2.0,>=1.17.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading yarl-1.18.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (67 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 67.7/67.7 kB 16.5 MB/s eta 0:00:00
+Requirement already satisfied: cryptography>=39.0 in /opt/conda/lib/python3.10/site-packages (from asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (42.0.8)
+Collecting azure-core<2.0.0,>=1.11.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading azure_core-1.32.0-py3-none-any.whl.metadata (39 kB)
+Collecting msal<2.0.0,>=1.12.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading msal-1.31.1-py3-none-any.whl.metadata (11 kB)
+Collecting msal-extensions<2.0.0,>=0.3.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading msal_extensions-1.2.0-py3-none-any.whl.metadata (7.6 kB)
+Collecting msrest>=0.6.21 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading msrest-0.7.1-py3-none-any.whl.metadata (21 kB)
+Collecting azure-common~=1.1 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading azure_common-1.1.28-py2.py3-none-any.whl.metadata (5.0 kB)
+Collecting azure-mgmt-core<2.0.0,>=1.3.1 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading azure_mgmt_core-1.5.0-py3-none-any.whl.metadata (4.3 kB)
+Collecting python-dateutil>=2.7.5 (from hcloud~=1.17->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl.metadata (8.4 kB)
+Requirement already satisfied: MarkupSafe>=2.0 in /opt/conda/lib/python3.10/site-packages (from jinja2>=2.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.1.5)
+INFO: pip is looking at multiple versions of jsonschema to determine which version is compatible with other requirements. This could take a while.
+Collecting jsonschema>=3.0.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading jsonschema-4.22.0-py3-none-any.whl.metadata (8.2 kB)
+  Downloading jsonschema-4.21.1-py3-none-any.whl.metadata (7.8 kB)
+  Downloading jsonschema-4.21.0-py3-none-any.whl.metadata (8.0 kB)
+  Downloading jsonschema-4.20.0-py3-none-any.whl.metadata (8.1 kB)
+  Downloading jsonschema-4.19.2-py3-none-any.whl.metadata (7.9 kB)
+  Downloading jsonschema-4.19.1-py3-none-any.whl.metadata (7.9 kB)
+  Downloading jsonschema-4.19.0-py3-none-any.whl.metadata (8.2 kB)
+INFO: pip is still looking at multiple versions of jsonschema to determine which version is compatible with other requirements. This could take a while.
+  Downloading jsonschema-4.18.6-py3-none-any.whl.metadata (7.8 kB)
+  Downloading jsonschema-4.18.5-py3-none-any.whl.metadata (7.7 kB)
+  Downloading jsonschema-4.18.4-py3-none-any.whl.metadata (7.8 kB)
+  Downloading jsonschema-4.18.3-py3-none-any.whl.metadata (7.9 kB)
+  Downloading jsonschema-4.18.2-py3-none-any.whl.metadata (7.8 kB)
+INFO: This is taking longer than usual. You might need to provide the dependency resolver with stricter constraints to reduce runtime. See https://pip.pypa.io/warnings/backtracking for guidance. If you want to abort this run, press Ctrl + C.
+  Downloading jsonschema-4.18.1-py3-none-any.whl.metadata (7.8 kB)
+  Downloading jsonschema-4.18.0-py3-none-any.whl.metadata (10 kB)
+  Downloading jsonschema-4.17.3-py3-none-any.whl.metadata (7.9 kB)
+Collecting pyrsistent!=0.17.0,!=0.17.1,!=0.17.2,>=0.14.0 (from jsonschema>=3.0.1->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading pyrsistent-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (27 kB)
+Collecting contourpy>=1.0.1 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)
+  Downloading contourpy-1.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.4 kB)
+Collecting cycler>=0.10 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)
+  Downloading cycler-0.12.1-py3-none-any.whl.metadata (3.8 kB)
+Collecting fonttools>=4.22.0 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)
+  Downloading fonttools-4.55.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (164 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 164.5/164.5 kB 13.7 MB/s eta 0:00:00
+Collecting kiwisolver>=1.3.1 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)
+  Downloading kiwisolver-1.4.7-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl.metadata (6.3 kB)
+Collecting pillow>=8 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)
+  Downloading pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl.metadata (9.1 kB)
+Collecting httplib2 (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading httplib2-0.22.0-py3-none-any.whl.metadata (2.6 kB)
+Collecting ujson (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading ujson-5.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (9.3 kB)
+Collecting pandas (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (89 kB)
+     ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 89.9/89.9 kB 29.1 MB/s eta 0:00:00
+Collecting jmespath (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading jmespath-1.0.1-py3-none-any.whl.metadata (7.6 kB)
+Collecting scramp>=1.4.5 (from pg8000~=1.19->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading scramp-1.4.5-py3-none-any.whl.metadata (19 kB)
+Collecting bs4 (from pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading bs4-0.0.2-py2.py3-none-any.whl.metadata (411 bytes)
+Collecting docutils (from python-daemon~=2.3->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading docutils-0.21.2-py3-none-any.whl.metadata (2.8 kB)
+Collecting lockfile>=0.10 (from python-daemon~=2.3->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading lockfile-0.12.2-py2.py3-none-any.whl.metadata (2.4 kB)
+Collecting imbalanced-learn (from imblearn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading imbalanced_learn-0.12.4-py3-none-any.whl.metadata (8.3 kB)
+Collecting joblib>=1.2.0 (from scikit-learn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading joblib-1.4.2-py3-none-any.whl.metadata (5.4 kB)
+Collecting threadpoolctl>=3.1.0 (from scikit-learn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)
+  Downloading threadpoolctl-3.5.0-py3-none-any.whl.metadata (13 kB)
+Requirement already satisfied: aiormq<4,>=3.2.3 in /opt/conda/lib/python3.10/site-packages (from aio-pika~=6.6->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.3.1)
+Requirement already satisfied: Mako in /opt/conda/lib/python3.10/site-packages (from alembic~=1.2->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.3.5)
+Requirement already satisfied: pyzmq>=17.0 in /opt/conda/lib/python3.10/site-packages (from circus~=0.18.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (26.0.3)
+Requirement already satisfied: tornado>=5.0.2 in /opt/conda/lib/python3.10/site-packages (from circus~=0.18.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.4)
+Requirement already satisfied: cffi>=1.12 in /opt/conda/lib/python3.10/site-packages (from cryptography>=39.0->asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (1.16.0)
+Requirement already satisfied: zipp>=0.5 in /opt/conda/lib/python3.10/site-packages (from importlib-metadata~=6.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.19.2)
+Requirement already satisfied: decorator in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.1.1)
+Requirement already satisfied: matplotlib-inline in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.1.7)
+Requirement already satisfied: prompt-toolkit<3.1.0,>=3.0.41 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.0.46)
+Requirement already satisfied: pygments>=2.4.0 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.18.0)
+Requirement already satisfied: stack-data in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.6.3)
+Requirement already satisfied: traitlets>=5.13.0 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.14.3)
+Requirement already satisfied: exceptiongroup in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.2.1)
+Requirement already satisfied: pexpect>4.3 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.9.0)
+Requirement already satisfied: parso<0.9.0,>=0.8.0 in /opt/conda/lib/python3.10/site-packages (from jedi<0.19->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.8.4)
+Requirement already satisfied: shortuuid in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.0.13)
+Requirement already satisfied: async-generator in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.10)
+Requirement already satisfied: pytray<0.4.0,>=0.2.2 in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.3.4)
+Requirement already satisfied: deprecation in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.1.0)
+Requirement already satisfied: pamqp~=2.0 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.3.0)
+Collecting PyJWT<3,>=1.0.0 (from PyJWT[crypto]<3,>=1.0.0->msal<2.0.0,>=1.12.0->azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading PyJWT-2.10.0-py3-none-any.whl.metadata (4.0 kB)
+Collecting portalocker<3,>=1.4 (from msal-extensions<2.0.0,>=0.3.0->azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading portalocker-2.10.1-py3-none-any.whl.metadata (8.5 kB)
+Requirement already satisfied: certifi>=2017.4.17 in /opt/conda/lib/python3.10/site-packages (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1) (2024.2.2)
+Collecting isodate>=0.6.0 (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading isodate-0.7.2-py3-none-any.whl.metadata (11 kB)
+Collecting requests-oauthlib>=0.5.0 (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading requests_oauthlib-2.0.0-py2.py3-none-any.whl.metadata (11 kB)
+Requirement already satisfied: bcrypt>=3.1.3 in /opt/conda/lib/python3.10/site-packages (from paramiko>=2.7.2,~=2.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.1.3)
+Requirement already satisfied: pynacl>=1.0.1 in /opt/conda/lib/python3.10/site-packages (from paramiko>=2.7.2,~=2.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.5.0)
+Requirement already satisfied: nest_asyncio~=1.5 in /opt/conda/lib/python3.10/site-packages (from plumpy~=0.21.6->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.6.0)
+Requirement already satisfied: annotated-types>=0.4.0 in /opt/conda/lib/python3.10/site-packages (from pydantic~=2.4->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.0)
+Requirement already satisfied: pydantic-core==2.18.4 in /opt/conda/lib/python3.10/site-packages (from pydantic~=2.4->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.18.4)
+Requirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.3.2)
+Requirement already satisfied: idna<4,>=2.5 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.7)
+Requirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.2.1)
+Collecting asn1crypto>=1.5.1 (from scramp>=1.4.5->pg8000~=1.19->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading asn1crypto-1.5.1-py2.py3-none-any.whl.metadata (13 kB)
+Requirement already satisfied: greenlet!=0.4.17 in /opt/conda/lib/python3.10/site-packages (from sqlalchemy~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.0.3)
+Collecting beautifulsoup4 (from bs4->pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading beautifulsoup4-4.12.3-py3-none-any.whl.metadata (3.8 kB)
+Collecting tzdata>=2022.7 (from pandas->mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading tzdata-2024.2-py2.py3-none-any.whl.metadata (1.4 kB)
+Requirement already satisfied: pycparser in /opt/conda/lib/python3.10/site-packages (from cffi>=1.12->cryptography>=39.0->asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (2.22)
+Requirement already satisfied: ptyprocess>=0.5 in /opt/conda/lib/python3.10/site-packages (from pexpect>4.3->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.0)
+Requirement already satisfied: wcwidth in /opt/conda/lib/python3.10/site-packages (from prompt-toolkit<3.1.0,>=3.0.41->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.13)
+Collecting oauthlib>=3.0.0 (from requests-oauthlib>=0.5.0->msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)
+  Downloading oauthlib-3.2.2-py3-none-any.whl.metadata (7.5 kB)
+Collecting soupsieve>1.2 (from beautifulsoup4->bs4->pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)
+  Downloading soupsieve-2.6-py3-none-any.whl.metadata (4.6 kB)
+Requirement already satisfied: executing>=1.2.0 in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.0.1)
+Requirement already satisfied: asttokens>=2.1.0 in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.4.1)
+Requirement already satisfied: pure-eval in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.2)
+Downloading ase-3.23.0-py3-none-any.whl (2.9 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 2.9/2.9 MB 74.6 MB/s eta 0:00:00
+Using cached reentry-1.3.3-py3-none-any.whl (17 kB)
+Downloading yascheduler-1.2.0-py3-none-any.whl (67 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 67.9/67.9 kB 21.7 MB/s eta 0:00:00
+Downloading spglib-2.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (515 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 515.3/515.3 kB 76.2 MB/s eta 0:00:00
+Downloading aiohttp-3.11.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.6 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.6/1.6 MB 105.2 MB/s eta 0:00:00
+Downloading asyncssh-2.18.0-py3-none-any.whl (367 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 367.6/367.6 kB 63.8 MB/s eta 0:00:00
+Downloading asyncstdlib-3.13.0-py3-none-any.whl (43 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 43.9/43.9 kB 10.0 MB/s eta 0:00:00
+Downloading attrs-21.4.0-py2.py3-none-any.whl (60 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 60.6/60.6 kB 18.9 MB/s eta 0:00:00
+Downloading azure_identity-1.10.0-py3-none-any.whl (134 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 134.1/134.1 kB 36.8 MB/s eta 0:00:00
+Downloading azure_mgmt_compute-27.2.0-py3-none-any.whl (5.7 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 5.7/5.7 MB 47.0 MB/s eta 0:00:00
+Downloading azure_mgmt_network-20.0.0-py3-none-any.whl (8.5 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.5/8.5 MB 115.1 MB/s eta 0:00:00
+Downloading backoff-2.1.2-py3-none-any.whl (14 kB)
+Downloading hcloud-1.35.0-py3-none-any.whl (85 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 85.4/85.4 kB 21.5 MB/s eta 0:00:00
+Downloading jsonschema-4.17.3-py3-none-any.whl (90 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 90.4/90.4 kB 27.0 MB/s eta 0:00:00
+Downloading matplotlib-3.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.3 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 8.3/8.3 MB 117.1 MB/s eta 0:00:00
+Downloading pg8000-1.31.2-py3-none-any.whl (54 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 54.5/54.5 kB 16.4 MB/s eta 0:00:00
+Downloading pyparsing-3.2.0-py3-none-any.whl (106 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 106.9/106.9 kB 31.0 MB/s eta 0:00:00
+Downloading python_daemon-2.3.2-py3-none-any.whl (34 kB)
+Downloading scipy-1.14.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (41.2 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 41.2/41.2 MB 69.5 MB/s eta 0:00:00
+Downloading upcloud_api-2.6.0-py3-none-any.whl (37 kB)
+Downloading imblearn-0.0-py2.py3-none-any.whl (1.9 kB)
+Downloading scikit_learn-1.5.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.3 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 13.3/13.3 MB 118.0 MB/s eta 0:00:00
+Downloading aiohappyeyeballs-2.4.3-py3-none-any.whl (14 kB)
+Downloading aiosignal-1.3.1-py3-none-any.whl (7.6 kB)
+Downloading async_timeout-5.0.1-py3-none-any.whl (6.2 kB)
+Downloading azure_common-1.1.28-py2.py3-none-any.whl (14 kB)
+Downloading azure_core-1.32.0-py3-none-any.whl (198 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 198.9/198.9 kB 47.3 MB/s eta 0:00:00
+Downloading azure_mgmt_core-1.5.0-py3-none-any.whl (30 kB)
+Downloading contourpy-1.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (324 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 325.0/325.0 kB 65.3 MB/s eta 0:00:00
+Downloading cycler-0.12.1-py3-none-any.whl (8.3 kB)
+Downloading fonttools-4.55.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (4.6 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.6/4.6 MB 113.3 MB/s eta 0:00:00
+Downloading frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (241 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 241.9/241.9 kB 51.4 MB/s eta 0:00:00
+Downloading joblib-1.4.2-py3-none-any.whl (301 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 301.8/301.8 kB 62.3 MB/s eta 0:00:00
+Downloading kiwisolver-1.4.7-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (1.6 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 1.6/1.6 MB 102.4 MB/s eta 0:00:00
+Downloading lockfile-0.12.2-py2.py3-none-any.whl (13 kB)
+Downloading msal-1.31.1-py3-none-any.whl (113 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 113.2/113.2 kB 30.9 MB/s eta 0:00:00
+Downloading msal_extensions-1.2.0-py3-none-any.whl (19 kB)
+Downloading msrest-0.7.1-py3-none-any.whl (85 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 85.4/85.4 kB 19.9 MB/s eta 0:00:00
+Downloading pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl (4.4 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 4.4/4.4 MB 118.7 MB/s eta 0:00:00
+Downloading propcache-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (208 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 208.9/208.9 kB 36.7 MB/s eta 0:00:00
+Downloading pyrsistent-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (117 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 117.7/117.7 kB 27.4 MB/s eta 0:00:00
+Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 229.9/229.9 kB 42.9 MB/s eta 0:00:00
+Downloading scramp-1.4.5-py3-none-any.whl (12 kB)
+Downloading threadpoolctl-3.5.0-py3-none-any.whl (18 kB)
+Downloading yarl-1.18.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (319 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 319.4/319.4 kB 67.3 MB/s eta 0:00:00
+Downloading bs4-0.0.2-py2.py3-none-any.whl (1.2 kB)
+Downloading docutils-0.21.2-py3-none-any.whl (587 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 587.4/587.4 kB 43.1 MB/s eta 0:00:00
+Downloading httplib2-0.22.0-py3-none-any.whl (96 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 96.9/96.9 kB 26.1 MB/s eta 0:00:00
+Downloading imbalanced_learn-0.12.4-py3-none-any.whl (258 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 258.3/258.3 kB 47.1 MB/s eta 0:00:00
+Downloading jmespath-1.0.1-py3-none-any.whl (20 kB)
+Downloading pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.1 MB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 13.1/13.1 MB 118.5 MB/s eta 0:00:00
+Downloading ujson-5.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (53 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 53.6/53.6 kB 15.2 MB/s eta 0:00:00
+Downloading asn1crypto-1.5.1-py2.py3-none-any.whl (105 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 105.0/105.0 kB 31.3 MB/s eta 0:00:00
+Downloading isodate-0.7.2-py3-none-any.whl (22 kB)
+Downloading portalocker-2.10.1-py3-none-any.whl (18 kB)
+Downloading PyJWT-2.10.0-py3-none-any.whl (23 kB)
+Downloading requests_oauthlib-2.0.0-py2.py3-none-any.whl (24 kB)
+Downloading tzdata-2024.2-py2.py3-none-any.whl (346 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 346.6/346.6 kB 62.6 MB/s eta 0:00:00
+Downloading beautifulsoup4-4.12.3-py3-none-any.whl (147 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 147.9/147.9 kB 38.3 MB/s eta 0:00:00
+Downloading oauthlib-3.2.2-py3-none-any.whl (151 kB)
+   ━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━━ 151.7/151.7 kB 42.8 MB/s eta 0:00:00
+Downloading soupsieve-2.6-py3-none-any.whl (36 kB)
+Building wheels for collected packages: mpds-aiida, aiida-crystal-dft, mpds-ml-labs, mpds_client, pycrystal, progressbar, pycodcif
+  Building wheel for mpds-aiida (setup.py): started
+  Building wheel for mpds-aiida (setup.py): finished with status 'done'
+  Created wheel for mpds-aiida: filename=mpds_aiida-0.10.1-py3-none-any.whl size=26528 sha256=425ccfd63a745288db9c902e668f77529951df66db7a8f44321478923363b091
+  Stored in directory: /tmp/pip-ephem-wheel-cache-erc7izfh/wheels/78/53/54/1f9951ffdf9d9f80855f9b673b98a29c0fd5474eda71dda1dd
+  Building wheel for aiida-crystal-dft (pyproject.toml): started
+  Building wheel for aiida-crystal-dft (pyproject.toml): finished with status 'done'
+  Created wheel for aiida-crystal-dft: filename=aiida_crystal_dft-0.9.1-py3-none-any.whl size=615109 sha256=823d3da78287542b4ac60ffca9a7448e9e38fc571591562ced16f8690daf6846
+  Stored in directory: /tmp/pip-ephem-wheel-cache-erc7izfh/wheels/a3/99/76/53856ab4749a880e808f4f2a1fa0f13576e6bef968abb3337c
+  Building wheel for mpds-ml-labs (setup.py): started
+  Building wheel for mpds-ml-labs (setup.py): finished with status 'done'
+  Created wheel for mpds-ml-labs: filename=mpds_ml_labs-0.0.7-py3-none-any.whl size=36850 sha256=6f681fe74e2df6e644b1d13721fa2dc241f6c2616eafe0438641dbfca9a82b5a
+  Stored in directory: /tmp/pip-ephem-wheel-cache-erc7izfh/wheels/4c/ed/d3/e721adc8f33103079f4f5fae1381741cd3f5a08d09cf18e59b
+  Building wheel for mpds_client (setup.py): started
+  Building wheel for mpds_client (setup.py): finished with status 'done'
+  Created wheel for mpds_client: filename=mpds_client-0.24-py3-none-any.whl size=9969 sha256=66dd9efb0add0ec83f4c18ffd23e127450f6c8a03ca33e7b154288e51dda88a3
+  Stored in directory: /home/aiida/.cache/pip/wheels/ac/3b/35/86c36e5a6497d198803077160a5ef38ef680c3a20750d01cdb
+  Building wheel for pycrystal (setup.py): started
+  Building wheel for pycrystal (setup.py): finished with status 'done'
+  Created wheel for pycrystal: filename=pycrystal-1.0.16-py3-none-any.whl size=27471 sha256=ef41fccfee9ffab7a3812b4aca11fa0b34b09243ee8b1bd553a5533b7a820036
+  Stored in directory: /home/aiida/.cache/pip/wheels/bb/b8/bd/373832f07727578241660773b3994e17776c78ca6a241f3362
+  Building wheel for progressbar (setup.py): started
+  Building wheel for progressbar (setup.py): finished with status 'done'
+  Created wheel for progressbar: filename=progressbar-2.5-py3-none-any.whl size=12067 sha256=2df91d6baab15b8717135f02873c6d5691312ddc9800fef4d866dc9330e5add3
+  Stored in directory: /home/aiida/.cache/pip/wheels/cd/17/e5/765d1a3112ff3978f70223502f6047e06c43a24d7c5f8ff95b
+  Building wheel for pycodcif (setup.py): started
+  Building wheel for pycodcif (setup.py): finished with status 'error'
+  error: subprocess-exited-with-error
+  
+  × python setup.py bdist_wheel did not run successfully.
+  │ exit code: 1
+  ╰─> [8 lines of output]
+      running bdist_wheel
+      running build
+      running build_py
+      running build_ext
+      building 'pycodcif._pycodcif' extension
+      swigging pycodcif.i to pycodcif_wrap.c
+      swig -python -o pycodcif_wrap.c pycodcif.i
+      error: command 'swig' failed: No such file or directory
+      [end of output]
+  
+  note: This error originates from a subprocess, and is likely not a problem with pip.
+  ERROR: Failed building wheel for pycodcif
+  Running setup.py clean for pycodcif
+Successfully built mpds-aiida aiida-crystal-dft mpds-ml-labs mpds_client pycrystal progressbar
+Failed to build pycodcif
+ERROR: Could not build wheels for pycodcif, which is required to install pyproject.toml-based projects
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'crystal.mpds' does not start with prefix 'mpds.'","W010: Entry point 'crystal.cif' does not start with prefix 'mpds.'","W010: Entry point 'crystal.aiida' does not start with prefix 'mpds.'"],summaryinfo:[{colorclass:"green",text:"Workflows",count:3}],pip_install_cmd:"pip install git+https://github.com/mpds-io/mpds-aiida",is_installable:"False"},"aiida-muon":{entry_point_prefix:"muon",code_home:"https://github.com/positivemuon/aiida-muon",version_file:"https://raw.githubusercontent.com/positivemuon/aiida-muon/main/aiida_muon/__init__.py",pip_url:"git+https://github.com/positivemuon/aiida-muon",name:"aiida-muon",package_name:"aiida_muon",hosted_on:"github.com",metadata:{description:"aiida-muon is allows to find candiate muon implantation sites and hyperfine field by DFT supercell relaxations and from further symmetry and kinetics analysis. ",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: Free To Use But Restricted","Natural Language :: English","Development Status :: 5 - Production/Stable","Framework :: AiiDA"],author:"Muon group Parma"},aiida_version:">=2.0,<3",entry_points:{"aiida.workflows":{"muon.find_muon":{description:["FindMuonWorkChain finds the candidate implantation site for a positive muon."," It first performs DFT relaxation calculations for a set of initial muon sites."," It then analyzes the results of these calculations and finds candidate muon sites."," If there are magnetic inequivalent sites not initially, they are recalculated"," It further calculates the muon contact hyperfine field at these candidate sites."],spec:{inputs:[{name:"relax",required:!0,valid_types:"Data",info:"Inputs for SCF calculations."},{name:"charge_supercell",required:!1,valid_types:"Bool, NoneType",info:"To run charged supercell for positive muon or not (neutral supercell)"},{name:"hubbard",required:!1,valid_types:"Bool, NoneType",info:"To check and get Hubbard U value or not"},{name:"impuritysupercellconv",required:!1,valid_types:"Data",info:"the preprocess MusconvWorkChain step, if needed."},{name:"impuritysupercellconv_metadata",required:!1,valid_types:"dict, NoneType",info:" Preferred metadata and scheduler options for impuritysupercellconv"},{name:"kpoints_distance",required:!1,valid_types:"Float",info:"The minimum desired distance in 1/Å between k-points in reciprocal space."},{name:"mag_dict",required:!1,valid_types:"Dict, NoneType",info:"magnetic dict created in protocols."},{name:"magmom",required:!1,valid_types:"List, NoneType",info:"List of 3D magnetic moments in Bohr magneton of the corresponding input structure if magnetic"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"mu_spacing",required:!1,valid_types:"Float, NoneType",info:"Minimum distance in Angstrom between two starting muon positions generated on a grid."},{name:"pp_code",required:!1,valid_types:"Code, NoneType",info:"The pp.x code-computer for post processing only if magmom is supplied"},{name:"pp_metadata",required:!1,valid_types:"dict, NoneType",info:" Preferred metadata and scheduler options for pp.x"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:"The label of the pseudo family"},{name:"pwscf",required:!1,valid_types:"Data",info:"Inputs for final SCF calculation with the muon at the origin."},{name:"qe_settings",required:!1,valid_types:"Dict, NoneType",info:" Preferred settings for the calc, otherwise default is used"},{name:"sc_matrix",required:!1,valid_types:"List, NoneType",info:" List of length 1 for supercell size "},{name:"spin_pol_dft",required:!1,valid_types:"Bool, NoneType",info:"Spin-polarised DFT simulation or not"},{name:"structure",required:!1,valid_types:"StructureData, StructureData, NoneType",info:"Input initial structure"}],outputs:[{name:"all_index_uuid",required:!0,valid_types:"Dict",info:""},{name:"all_sites",required:!0,valid_types:"Dict",info:""},{name:"unique_sites",required:!0,valid_types:"Dict",info:""},{name:"unique_sites_dipolar",required:!1,valid_types:"List",info:""},{name:"unique_sites_hyperfine",required:!1,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:404,message:"The MusconvWorkChain subprocesses failed"},{status:405,message:"One of the PwRelaxWorkChain subprocesses failed"},{status:406,message:"One of the PwBaseWorkChain subprocesses failed"},{status:407,message:"One of the PPWorkChain subprocesses failed"},{status:407,message:"No supercells available: try to decrease mu_spacing."}]},class:"aiida_muon.workflows.find_muon:FindMuonWorkChain"}}},commits_count:84,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/positivemuon/aiida-muon",is_installable:"True"},"aiida-musconv":{entry_point_prefix:"musconv",code_home:"https://github.com/positivemuon/aiida-musconv",version_file:"raw.githubusercontent.com/positivemuon/aiida-musconv/main/aiida_musconv/__init__.py",pip_url:"git+https://github.com/positivemuon/aiida-musconv",name:"aiida-musconv",package_name:"aiida_musconv",hosted_on:"github.com",metadata:{description:"aiida-impuritysupercellconv is a plugin that allows to obtain converged supercell size for an interstitial impurity calculation.",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: Free To Use But Restricted","Natural Language :: English","Development Status :: 5 - Production/Stable","Framework :: AiiDA"],author:"Muon group Parma"},aiida_version:">=2.0,<3",entry_points:{"aiida.workflows":{impuritysupercellconv:"aiida_impuritysupercellconv.workflows.impuritysupercellconv:IsolatedImpurityWorkChain"}},commits_count:46,development_status:"stable",errors:[`E002: Failed to import package aiida_musconv
Traceback (most recent call last):
+  File "", line 1, in 
+ModuleNotFoundError: No module named 'aiida_musconv'
+
`],warnings:["W010: Entry point 'impuritysupercellconv' does not start with prefix 'musconv.'"],summaryinfo:[{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/positivemuon/aiida-musconv",is_installable:"True"},"aiida-nanotech-empa":{code_home:"https://github.com/nanotech-empa/aiida-nanotech-empa",development_status:"beta",entry_point_prefix:"nanotech_empa",pip_url:"git+https://github.com/nanotech-empa/aiida-nanotech-empa",name:"aiida-nanotech-empa",package_name:"aiida_nanotech_empa",hosted_on:"github.com",metadata:{description:"AiiDA plugins and workflows developed at nanotech@surfaces group from Empa.",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python","Programming Language :: Python :: 3.9","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Topic :: Scientific/Engineering"],author:"nanotech@surfaces"},aiida_version:">=2.4.3,<3.0.0",entry_points:{"aiida.parsers":{"nanotech_empa.pp":"aiida_nanotech_empa.parsers:PpParser","nanotech_empa.cp2k_gw_parser":"aiida_nanotech_empa.parsers:Cp2kGwParser","nanotech_empa.cp2k_neb_parser":"aiida_nanotech_empa.parsers:Cp2kNebParser","nanotech_empa.gaussian.cubegen_pymol":"aiida_nanotech_empa.parsers.cubegen_pymol_parser:CubegenPymolParser","nanotech_empa.gaussian.casscf":"aiida_nanotech_empa.parsers:GaussianCasscfParser"},"aiida.calculations":{"nanotech_empa.stm":{description:["This is a StmCalculation."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"STM input parameters"},{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"remote folder"},{name:"settings",required:!0,valid_types:"Dict",info:"special settings"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_nanotech_empa.plugins:StmCalculation"},"nanotech_empa.overlap":{description:["No description available"],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Overlap input parameters"},{name:"parent_mol_folder",required:!0,valid_types:"RemoteData",info:"molecule scf folder"},{name:"parent_slab_folder",required:!0,valid_types:"RemoteData",info:"slab scf folder"},{name:"settings",required:!0,valid_types:"Dict",info:"special settings"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_nanotech_empa.plugins:OverlapCalculation"},"nanotech_empa.afm":{description:["No description available"],spec:{inputs:[{name:"atomtypes",required:!0,valid_types:"SinglefileData",info:"atomtypes.ini file"},{name:"geo_no_labels",required:!0,valid_types:"SinglefileData",info:"geometry without spin labels file"},{name:"parameters",required:!0,valid_types:"Dict",info:"AFM input parameters"},{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"remote folder"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_nanotech_empa.plugins:AfmCalculation"},"nanotech_empa.hrstm":{description:["No description available"],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"HRSTM input parameters"},{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"scf folder"},{name:"ppm_calc_folder",required:!0,valid_types:"RemoteData",info:"ppm folder"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_nanotech_empa.plugins:HrstmCalculation"}},"aiida.workflows":{"nanotech_empa.nanoribbon":{description:["No description available"],spec:{inputs:[{name:"pp_code",required:!0,valid_types:"Code",info:""},{name:"projwfc_code",required:!0,valid_types:"Code",info:""},{name:"pseudo_family",required:!0,valid_types:"Str",info:"An alternative to specifying the pseudo potentials manually in `pseudos`: one can specify the name of an existing pseudo potential family and the work chain will generate the pseudos automatically based on the input structure."},{name:"pw_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"max_kpoints",required:!1,valid_types:"Int, NoneType",info:""},{name:"max_nodes",required:!1,valid_types:"Int, NoneType",info:""},{name:"mem_node",required:!1,valid_types:"Int, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"num_export_bands",required:!1,valid_types:"Int, NoneType",info:""},{name:"optimize_cell",required:!1,valid_types:"Bool, NoneType",info:""},{name:"precision",required:!1,valid_types:"Float, NoneType",info:""},{name:"tot_charge",required:!1,valid_types:"Float, NoneType",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed."}]},class:"aiida_nanotech_empa.workflows.qe:NanoribbonWorkChain"},"nanotech_empa.gaussian.scf":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"functional",required:!0,valid_types:"Str",info:"xc functional"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"cdiis",required:!1,valid_types:"Bool, NoneType",info:"Conjugate Direct Inversion in the Iterative Subspace"},{name:"conver",required:!1,valid_types:"Int, NoneType",info:"the scf convergece threshold"},{name:"cubegen_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"cubegen_parser_name",required:!1,valid_types:"str",info:""},{name:"cubegen_parser_params",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters to cubegen parser."},{name:"cubes_edge_space",required:!1,valid_types:"Float, NoneType",info:"Extra cube space in addition to bounding box [ang]."},{name:"cubes_n_occ",required:!1,valid_types:"Int, NoneType",info:"Number of occupied orbital cubes to generate"},{name:"cubes_n_virt",required:!1,valid_types:"Int, NoneType",info:"Number of virtual orbital cubes to generate"},{name:"empirical_dispersion",required:!1,valid_types:"Str, NoneType",info:'Include empirical dispersion corrections(e.g. "GD3", "GD3BJ")'},{name:"formchk_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"int",required:!1,valid_types:"Str, NoneType",info:"the integral grid"},{name:"maxcycle",required:!1,valid_types:"Int, NoneType",info:"the maximum number of scf cycles"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"Spin multiplicity; 0 means RKS"},{name:"nmr",required:!1,valid_types:"Bool, NoneType",info:"nmr calculation"},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"the folder of a completed gaussian calculation"},{name:"wfn_stable_opt",required:!1,valid_types:"Bool, NoneType",info:"If true, perform wfn stability optimization."},{name:"wfn_stable_opt_min_basis",required:!1,valid_types:"Bool, NoneType",info:"If true, perform first a minimal basis stability opt."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"Multiplicity and number of el. doesn't match."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianScfWorkChain"},"nanotech_empa.gaussian.relax":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"functional",required:!0,valid_types:"Str",info:"xc functional"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"basis_set_scf",required:!1,valid_types:"Str, NoneType",info:"Basis set for SCF. If not present, SCF is skipped."},{name:"cdiis",required:!1,valid_types:"Bool, NoneType",info:"Conjugate Direct Inversion in the Iterative Subspace"},{name:"constraints",required:!1,valid_types:"List, NoneType",info:'Supported constraints: ("distance", n1, n2, d)'},{name:"conver",required:!1,valid_types:"Int, NoneType",info:"the scf convergece threshold"},{name:"cubegen_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"cubegen_parser_name",required:!1,valid_types:"str",info:""},{name:"cubegen_parser_params",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters to cubegen parser."},{name:"cubes_edge_space",required:!1,valid_types:"Float, NoneType",info:"Extra cube space in addition to bounding box [ang]."},{name:"cubes_n_occ",required:!1,valid_types:"Int, NoneType",info:"Number of occupied orbital cubes to generate"},{name:"cubes_n_virt",required:!1,valid_types:"Int, NoneType",info:"Number of virtual orbital cubes to generate"},{name:"empirical_dispersion",required:!1,valid_types:"Str, NoneType",info:'Include empirical dispersion corrections(e.g. "GD3", "GD3BJ")'},{name:"formchk_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"freq",required:!1,valid_types:"Bool, NoneType",info:"Also run vibrational analysis."},{name:"int",required:!1,valid_types:"Str, NoneType",info:"the integral grid"},{name:"maxcycle",required:!1,valid_types:"Int, NoneType",info:"the maximum number of scf cycles"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"spin multiplicity; 0 means RKS"},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of automatic."},{name:"tight",required:!1,valid_types:"Bool, NoneType",info:"Use tight optimization criteria."},{name:"wfn_stable_opt",required:!1,valid_types:"Bool, NoneType",info:"if true, perform wfn stability optimization"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"Multiplicity and number of el. doesn't match."},{status:302,message:"Input options are invalid."},{status:380,message:"Vibrational analysis did not succeed."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianRelaxWorkChain"},"nanotech_empa.gaussian.delta_scf":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"functional",required:!0,valid_types:"Str",info:"xc functional"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"spin multiplicity; 0 means RKS"},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"the folder of a completed gaussian calculation"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"Multiplicity and number of el. doesn't match."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianDeltaScfWorkChain"},"nanotech_empa.gaussian.natorb":{description:["No description available"],spec:{inputs:[{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"parent Gaussian calculation directory"},{name:"parent_calc_params",required:!0,valid_types:"Dict",info:"parent Gaussian calculation output parameters"},{name:"cubegen_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"cubegen_parser_params",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters to cubegen parser."},{name:"edge_space",required:!1,valid_types:"Float, NoneType",info:"Extra cube space in addition to molecule bounding box [ang]."},{name:"formchk_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"num_natural_orbital_cubes",required:!1,valid_types:"Int, NoneType",info:"Generate cubes for SAVED natural orbitals (n*occ and n*virt)."},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"save_natorb_chk",required:!1,valid_types:"Bool, NoneType",info:"Save natural orbitals in the chk file.Can introduce errors for larger systems"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianNatOrbWorkChain"},"nanotech_empa.gaussian.spin":{description:["No description available"],spec:{inputs:[{name:"basis_set_opt",required:!0,valid_types:"Str",info:"basis_set for opt"},{name:"basis_set_scf",required:!0,valid_types:"Str",info:"basis_set for scf"},{name:"cubegen_code",required:!0,valid_types:"Code",info:""},{name:"formchk_code",required:!0,valid_types:"Code",info:""},{name:"functional",required:!0,valid_types:"Str",info:"xc functional"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"multiplicity_list",required:!0,valid_types:"List",info:"spin multiplicities"},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"empirical_dispersion",required:!1,valid_types:"Str, NoneType",info:'Include empirical dispersion corrections(e.g. "GD3", "GD3BJ")'},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianSpinWorkChain"},"nanotech_empa.gaussian.hf_mp2":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"cubegen_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"cubegen_parser_params",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters to cubegen parser."},{name:"edge_space",required:!1,valid_types:"Float, NoneType",info:"Extra cube space in addition to molecule bounding box [ang]."},{name:"formchk_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"spin multiplicity; 0 means RKS"},{name:"num_orbital_cubes",required:!1,valid_types:"Int, NoneType",info:"Generate cubes for the mp2 orbitals (n*occ and n*virt)."},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"the folder of a completed gaussian calculation"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"Multiplicity and number of el. doesn't match."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianHfMp2WorkChain"},"nanotech_empa.gaussian.constr_opt_chain":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"functional",required:!0,valid_types:"Str",info:"xc functional"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"basis_set_scf",required:!1,valid_types:"Str, NoneType",info:"basis_set for SCF"},{name:"empirical_dispersion",required:!1,valid_types:"Str, NoneType",info:'Include empirical dispersion corrections(e.g. "GD3", "GD3BJ")'},{name:"extra_scf_mults",required:!1,valid_types:"List, NoneType",info:"Extra multiplicites for the SCF"},{name:"list_of_constraints",required:!1,valid_types:"List, NoneType",info:'Supported constraints: ("distance", n1, n2, d)'},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"spin multiplicity; 0 means RKS"},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"tight",required:!1,valid_types:"Bool, NoneType",info:"Use tight optimization criteria."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianConstrOptChainWorkChain"},"nanotech_empa.gaussian.casscf":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"m",required:!0,valid_types:"Int",info:"Number of orbitals CAS(n,m)."},{name:"n",required:!0,valid_types:"Int",info:"Number of electrons CAS(n,m)."},{name:"parent_calc_folder",required:!0,valid_types:"RemoteData",info:"parent Gaussian calculation directory"},{name:"cubegen_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"formchk_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"mp2",required:!1,valid_types:"Bool, NoneType",info:"calculate the MP2 correction (CASMP2)."},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"spin multiplicity"},{name:"num_orbital_cubes",required:!1,valid_types:"Int, NoneType",info:"Generate cubes for orbitals (n*occ and n*virt)."},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"uno",required:!1,valid_types:"Bool, NoneType",info:"Use the natural orbitals from the previous calculation."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianCasscfWorkChain"},"nanotech_empa.gaussian.casscf_series":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"Basis set"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"nm_list",required:!0,valid_types:"List",info:"Successive list of (n,m) tuples to run CAS(n,m)."},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"cubegen_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"formchk_code",required:!1,valid_types:"Code, NoneType",info:""},{name:"init_functional",required:!1,valid_types:"Str, NoneType",info:"Functional for the initial orbitals."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"mp2",required:!1,valid_types:"Bool, NoneType",info:"calculate the MP2 correction (CASMP2)."},{name:"multiplicity_list",required:!1,valid_types:"List, NoneType",info:"spin multiplicity"},{name:"num_orbital_cubes",required:!1,valid_types:"Int, NoneType",info:"Generate cubes for orbitals (n*occ and n*virt)."},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"start_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Read starting orbitals from here instead."},{name:"start_uno",required:!1,valid_types:"Bool, NoneType",info:"Use natural orbitals of the start calculation."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:302,message:"Input options are invalid."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianCasscfSeriesWorkChain"},"nanotech_empa.gaussian.nics":{description:["No description available"],spec:{inputs:[{name:"basis_set",required:!0,valid_types:"Str",info:"basis_set"},{name:"functional",required:!0,valid_types:"Str",info:"xc functional"},{name:"gaussian_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"input geometry"},{name:"empirical_dispersion",required:!1,valid_types:"Str, NoneType",info:'Include empirical dispersion corrections(e.g. "GD3", "GD3BJ")'},{name:"height",required:!1,valid_types:"Float, NoneType",info:"Height of NICS centers"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:"spin multiplicity; 0 means RKS"},{name:"opt",required:!1,valid_types:"Bool, NoneType",info:"False do not optimize structure"},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Use custom metadata.options instead of the automatic ones."},{name:"wfn_stable_opt",required:!1,valid_types:"Bool, NoneType",info:"if true, perform wfn stability optimization"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.gaussian:GaussianNicsWorkChain"},"nanotech_empa.cp2k.geo_opt":{description:["No description available"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"sys_params",required:!0,valid_types:"Dict",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the work chain (geo_opt_protocol)."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kGeoOptWorkChain"},"nanotech_empa.cp2k.fragment_separation":{description:["WorkChain to compute adsoprtion energy for a molecule on a substrate."],spec:{inputs:[{name:"auxilary_dictionaries",required:!0,valid_types:"Dict",info:"Dictionaries to append to the CP2K input dictionary right before the job submission. It is useful to add constraints and collective variables to the input."},{name:"code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:` + multiplicities, dictionary: multiplicity of each fragment. Use 'all' to specify the multiplicity of the whole system. + magnetization_per_site: Magnetization per site of the whole system. + Magnetization per site of the fragments will be extracted automatically. + charges, dictionary: Charges of each fragment. No need to specify the charge of the full system + as it would be computed automatically. + uks: Use unrestricted Kohn-Sham. + `},{name:"fragments",required:!0,valid_types:"List",info:"List of indices of atoms defining individual fragments."},{name:"structure",required:!0,valid_types:"StructureData",info:"A molecule on a substrate."},{name:"fixed_atoms",required:!1,valid_types:"List, NoneType",info:"Fixed atoms of the whole system. Fixed atoms of the fragments will be extracted automatically."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"int",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol used by the work chain (geo_opt_protocol)."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kFragmentSeparationWorkChain"},"nanotech_empa.cp2k.ads_gw_ic":{description:["WorkChain to run GW and IC for an adsorbed system",""," Two different ways to run:"," 1) geometry of a molecule adsorbed on a substrate"," 2) isolated molecule & adsorption height"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"A molecule on a substrate or an isolated molecule."},{name:"ads_height",required:!1,valid_types:"Float, NoneType",info:"Adsoprtion height from the molecular geometrical center.Required if an isolated molecule is specified."},{name:"geometry_mode",required:!1,valid_types:"Str, NoneType",info:"Possibilities: ads_geo, gas_opt"},{name:"magnetization_per_site",required:!1,valid_types:"List, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:""},{name:"options",required:!1,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kMoleculeGwWorkChain."},{name:"substrate",required:!1,valid_types:"Str",info:"Substrate type, determines the image charge plane."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:380,message:"Specified substrate is not supported."},{status:381,message:"Structure analysis failed."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kAdsorbedGwIcWorkChain"},"nanotech_empa.cp2k.molecule_gw":{description:["No description available"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"debug",required:!1,valid_types:"Bool, NoneType",info:"Run with fast parameters for debugging."},{name:"magnetization_per_site",required:!1,valid_types:"List, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:""},{name:"options",required:!1,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Either 'gapw_std', 'gapw_hq', 'gpw_std'"},{name:"run_image_charge",required:!1,valid_types:"Bool, NoneType",info:"Run the image charge correction calculation."},{name:"z_ic_plane",required:!1,valid_types:"Float, NoneType",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:381,message:"SCF of the first step did not converge."},{status:382,message:"SCF of the second step did not converge."},{status:383,message:"SCF produced a negative gap."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kMoleculeGwWorkChain"},"nanotech_empa.cp2k.mol_opt_gw":{description:["WorkChain to optimize molecule and run GW:",""," Two different ways to run:"," 1) optimize geo and run gw"," 2) run gw"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"An isolated molecule."},{name:"debug",required:!1,valid_types:"Bool, NoneType",info:"Run with fast parameters for debugging."},{name:"geo_opt",required:!1,valid_types:"Bool, NoneType",info:"Perform geo opt step."},{name:"magnetization_per_site",required:!1,valid_types:"List, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"multiplicity",required:!1,valid_types:"Int, NoneType",info:""},{name:"options",required:!1,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the GW workchain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kMoleculeOptGwWorkChain"},"nanotech_empa.cp2k.pdos":{description:["No description available"],spec:{inputs:[{name:"cp2k_code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"molecule_indices",required:!0,valid_types:"List",info:""},{name:"options",required:!0,valid_types:"int",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"overlap_code",required:!0,valid_types:"Code",info:""},{name:"overlap_params",required:!0,valid_types:"Dict",info:""},{name:"pdos_lists",required:!0,valid_types:"List",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:"Coordinates of the whole system."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kDiagWorkChain workchain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kPdosWorkChain"},"nanotech_empa.cp2k.orbitals":{description:["No description available"],spec:{inputs:[{name:"cp2k_code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"spm_code",required:!0,valid_types:"Code",info:""},{name:"spm_params",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kDiagWorkChain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kOrbitalsWorkChain"},"nanotech_empa.cp2k.stm":{description:["No description available"],spec:{inputs:[{name:"cp2k_code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"spm_code",required:!0,valid_types:"Code",info:""},{name:"spm_params",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kDiagWorkChain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kStmWorkChain"},"nanotech_empa.cp2k.afm":{description:["No description available"],spec:{inputs:[{name:"afm_2pp_code",required:!0,valid_types:"Code",info:""},{name:"afm_2pp_params",required:!0,valid_types:"Dict",info:""},{name:"afm_pp_code",required:!0,valid_types:"Code",info:""},{name:"afm_pp_params",required:!0,valid_types:"Dict",info:""},{name:"cp2k_code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kDiagWorkChain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kAfmWorkChain"},"nanotech_empa.cp2k.hrstm":{description:["No description available"],spec:{inputs:[{name:"cp2k_code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"hrstm_code",required:!0,valid_types:"Code",info:""},{name:"hrstm_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"ppm_code",required:!0,valid_types:"Code",info:""},{name:"ppm_params",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kDiagWorkChain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kHrstmWorkChain"},"nanotech_empa.cp2k.diag":{description:["No description available"],spec:{inputs:[{name:"cp2k_code",required:!0,valid_types:"Code",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"dft_params",required:!1,valid_types:"Dict",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"pdos_lists",required:!1,valid_types:"List, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kBaseWorkChain."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kDiagWorkChain"},"nanotech_empa.cp2k.replica":{description:["Workflow to run Replica Chain calculations with CP2K."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"sys_params",required:!0,valid_types:"Dict",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kBaseWorkChain."},{name:"restart_from",required:!1,valid_types:"Str, NoneType",info:""}],outputs:[{name:"details",required:!0,valid_types:"Dict",info:""},{name:"structures",required:!0,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One geo opt failed"}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kReplicaWorkChain"},"nanotech_empa.cp2k.neb":{description:["No description available"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"neb_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"sys_params",required:!0,valid_types:"Dict",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kDiagWorkChain."},{name:"replicas",required:!1,valid_types:"StructureData",info:"nodes of input replicas"},{name:"restart_from",required:!1,valid_types:"Str, NoneType",info:""},{name:"wfn_cp_commands",required:!1,valid_types:"Str, NoneType",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:380,message:"no structures specified"},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kNebWorkChain"},"nanotech_empa.cp2k.phonons":{description:["No description available"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"phonons_params",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"sys_params",required:!0,valid_types:"Dict",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kGeoOptWorkChain."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps of the work chain failed."}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kPhononsWorkChain"},"nanotech_empa.cp2k.reftraj":{description:["Workflow to run Replica Chain calculations with CP2K."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dft_params",required:!0,valid_types:"Dict",info:""},{name:"options",required:!0,valid_types:"dict",info:"Define options for the cacluations: walltime, memory, CPUs, etc."},{name:"sys_params",required:!0,valid_types:"Dict",info:""},{name:"trajectory",required:!0,valid_types:"TrajectoryData",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"num_batches",required:!1,valid_types:"Int",info:""},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:""},{name:"protocol",required:!1,valid_types:"Str, NoneType",info:"Protocol supported by the Cp2kBaseWorkChain."},{name:"restart",required:!1,valid_types:"Bool, NoneType",info:""}],outputs:[{name:"details",required:!0,valid_types:"Dict",info:""},{name:"structures",required:!0,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:390,message:"One or more steps failed"}]},class:"aiida_nanotech_empa.workflows.cp2k:Cp2kRefTrajWorkChain"}},"aiida.schedulers":{slurm_ethz_euler:"aiida_nanotech_empa.schedulers:ETHZEulerSlurmScheduler"}},commits_count:18,errors:[],warnings:["W005: Development status in classifiers (stable) does not match development_status in metadata (beta)","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'slurm_ethz_euler' does not start with prefix 'nanotech_empa.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:4},{colorclass:"brown",text:"Parsers",count:5},{colorclass:"green",text:"Workflows",count:26},{colorclass:"orange",text:"Other (Schedulers)",count:1}],pip_install_cmd:"pip install git+https://github.com/nanotech-empa/aiida-nanotech-empa",is_installable:"True"},"aiida-nims-scheduler":{code_home:"https://github.com/atztogo/aiida-nims-scheduler",development_status:"stable",documentation_url:"https://github.com/atztogo/aiida-nims-scheduler",entry_point_prefix:"nims_scheduler",pip_url:"git+https://github.com/atztogo/aiida-nims-scheduler",name:"aiida-nims-scheduler",package_name:"aiida_nims_scheduler",hosted_on:"github.com",metadata:{description:"NIMS supercomputer scheduler",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Development Status :: 3 - Alpha","Framework :: AiiDA"],author:"Atsushi Togo"},aiida_version:">=2.0,<3",entry_points:{"aiida.schedulers":{nims_scheduler:"aiida_nims_scheduler.schedulers.pbs_nims:PbsNimsScheduler"}},commits_count:34,errors:[],warnings:["W005: Development status in classifiers (alpha) does not match development_status in metadata (stable)","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"orange",text:"Other (Schedulers)",count:1}],pip_install_cmd:"pip install git+https://github.com/atztogo/aiida-nims-scheduler",is_installable:"True"},"aiida-nwchem":{code_home:"https://github.com/aiidateam/aiida-nwchem",documentation_url:"https://aiida-nwchem.readthedocs.io/",entry_point_prefix:"nwchem",pip_url:"aiida-nwchem",plugin_info:"https://raw.githubusercontent.com/aiidateam/aiida-nwchem/master/setup.json",name:"aiida-nwchem",package_name:"aiida_nwchem",hosted_on:"github.com",metadata:{release_date:"2023-08-22",description:"The official AiiDA plugin for NWChem",author_email:"The AiiDA team ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering"],version:"3.0.1"},aiida_version:">=2.0,<3.0",entry_points:{"aiida.calculations":{"nwchem.base":{description:["Base calculation class for NWChem."],spec:{inputs:[{name:"input_file",required:!0,valid_types:"SinglefileData",info:"NWChem input file"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_folder",required:!1,valid_types:"RemoteData, FolderData, NoneType",info:"Remote directory of a completed NWChem calculation to restart from."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Required output files are missing."},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"The stdout output file could not be read."},{status:312,message:"The stdout output file was incomplete."},{status:313,message:"The stdout contains multiple calculations"},{status:340,message:"The calculation stopped prematurely because it ran out of walltime but the job was killed by the scheduler before the files were safely written to disk for a potential restart."},{status:350,message:"The parser raised an unexpected exception."}]},class:"aiida_nwchem.calculations.nwchem:NwchemBaseCalculation"},"nwchem.nwchem":{description:["Base calculation class for NWChem.",""," Synthesizes NWChem input file from parameter dictionary and StructureData."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure, with or without a cell"},{name:"add_cell",required:!1,valid_types:"Bool",info:"The input structure, with or without a cell"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_folder",required:!1,valid_types:"RemoteData, FolderData, NoneType",info:"Remote directory of a completed NWChem calculation to restart from."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Required output files are missing."},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"The stdout output file could not be read."},{status:312,message:"The stdout output file was incomplete."},{status:313,message:"The stdout contains multiple calculations"},{status:340,message:"The calculation stopped prematurely because it ran out of walltime but the job was killed by the scheduler before the files were safely written to disk for a potential restart."},{status:350,message:"The parser raised an unexpected exception."}]},class:"aiida_nwchem.calculations.nwchem:NwchemCalculation"}},"aiida.parsers":{"nwchem.nwchem":"aiida_nwchem.parsers.nwchem:NwchemBaseParser"},"aiida.workflows":{"nwchem.base":{description:["Workchain to run an NWChem calculation with automated error handling and restarts."],spec:{inputs:[{name:"nwchem",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The relaxed output structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_nwchem.workflows.base:NwchemBaseWorkChain"}}},commits_count:0,development_status:"beta",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-nwchem",is_installable:"True"},"aiida-open_circuit_voltage":{entry_point_prefix:"quantumespresso.ocv",code_home:"https://github.com/tsthakur/aiida-open_circuit_voltage",name:"aiida-open_circuit_voltage",package_name:"aiida_open_circuit_voltage",hosted_on:"github.com",metadata:{author:"Tushar Thakur",author_email:"tushar.thakur@epfl.ch",version:"0.2.0",description:"The AiiDA plugin to calculate ocv at various charge of states using QE",classifiers:["Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Development Status :: 4 - Beta","Natural Language :: English","Intended Audience :: Science/Research"]},aiida_version:">=1.1.0",entry_points:{"aiida.workflows":{"quantumespresso.ocv.ocvwc":"aiida_open_circuit_voltage.workflows.workchain:OCVWorkChain"}},commits_count:6,development_status:"beta",errors:[],warnings:["W009: Prefix 'quantumespresso.ocv' does not follow naming convention."],summaryinfo:[{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"Missing, see source code repository"},"aiida-optimize":{code_home:"https://github.com/greschd/aiida-optimize",documentation_url:"https://aiida-optimize.readthedocs.io",entry_point_prefix:"optimize",pip_url:"aiida-optimize",plugin_info:"https://raw.githubusercontent.com/greschd/aiida-optimize/master/setup.json",name:"aiida-optimize",package_name:"aiida_optimize",hosted_on:"github.com",metadata:{release_date:"2023-03-30",description:"AiiDA Plugin for running optimization algorithms.",author:"Dominik Gresch",author_email:"greschd@gmx.ch",license:"Apache 2.0",home_page:"https://aiida-optimize.readthedocs.io/",classifiers:["Development Status :: 5 - Production/Stable","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: Apache Software License","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Physics"],version:"1.0.2"},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.workflows":{"optimize.optimize":{description:["Runs an optimization procedure, given an optimization engine that defines the optimization"," algorithm, and a process which evaluates the function to be optimized."],spec:{inputs:[{name:"engine",required:!0,valid_types:"Str",info:"Engine that runs the optimization."},{name:"engine_kwargs",required:!0,valid_types:"Dict",info:"Keyword arguments passed to the optimization engine."},{name:"evaluate_process",required:!0,valid_types:"Str",info:"Process which produces the result to be optimized."},{name:"evaluate",required:!1,valid_types:"",info:"Inputs that are passed to all evaluation processes."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"optimal_process_output",required:!0,valid_types:"",info:"Output value of the optimal evaluation process."},{name:"optimal_process_uuid",required:!0,valid_types:"",info:"UUID of the optimal evaluation process."},{name:"engine_outputs",required:!1,valid_types:"",info:""},{name:"optimal_process_input",required:!1,valid_types:"",info:"Input value of the optimal evaluation process."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"Optimization failed because one of the evaluate processes did not finish ok."},{status:202,message:"Optimization failed because the engine did not finish ok."}]},class:"aiida_optimize._optimization_workchain:OptimizationWorkChain"},"optimize.wrappers.add_inputs":{description:["Wrapper workchain that takes inputs as keys and values and passes it"," on to a sub-process. This enables taking a process which was not"," designed to be used in optimization, and optimize with respect to"," some arbitrary input. Inputs which always remain the same can be"," specified in the ``inputs`` namespace, whereas the inputs to be"," optimized are given through the ``added_input_keys`` and"," ``added_input_values`` inputs.",""," The outputs of the wrapper workchain are the same as those of"," the wrapped process.","",' The "added" inputs can only be BaseType sub-classes, or'," attributes of a Dict. For each input, its port location is given",' in the "added_input_keys" input. For example, ``x.y`` would set'," the ``y`` input in the ``x`` namespace.",""," For cases where the input is a Dict attribute, the (possibly nested) attribute name is given after a colon. That means ``x:a.b`` would"," set the ``['a']['b']`` attribute of the ``Dict`` given in the ``x``"," input.",""," In cases where only a single input needs to be added, they can be"," specified directly instead of wrapped in a List."],spec:{inputs:[{name:"added_input_keys",required:!0,valid_types:"List, Str",info:"Specifies the location of each added input."},{name:"added_input_values",required:!0,valid_types:"List, BaseType",info:"Values of the added inputs to be passed into the sub-process."},{name:"sub_process",required:!0,valid_types:"Str",info:"The class of the process that should be wrapped."},{name:"inputs",required:!1,valid_types:"",info:"Inputs to be passed on to the sub-process."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"Workchain failed because the sub-process did not finish ok."}]},class:"aiida_optimize.wrappers._add_inputs:AddInputsWorkChain"},"optimize.wrappers.concatenate":{description:["Allows concatenating an arbitrary number of sub-processes.",""," A wrapper workchain that allows concatenating an arbitrary number"," of sub-processes. Outputs of one processes can be configured to"," be passed to the next one."],spec:{inputs:[{name:"output_input_mappings",required:!0,valid_types:"List",info:"Defines how inputs are passed between sub-processes. Each list entry entry has the form `((process_label_a, process_label_b), mapping)`, and defines outputs of process A to be passed to process B. The `mapping` values are dictionaries `{'output_name': 'input_name'}` giving the output name (in process A) and input name (in process B) for each value to pass."},{name:"process_inputs",required:!0,valid_types:"",info:"Inputs which are passed on to the sub-processes. The inputs should be grouped into a namespace identified by the process label."},{name:"process_labels",required:!0,valid_types:"List",info:"A list of pairs (label, process_name). The labels can be any string, the process_name needs to be loadable by `aiida_optimize.process_inputs.load_object`, and defines which process is being run."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"process_outputs",required:!0,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"Workchain failed because a sub-process failed."}]},class:"aiida_optimize.wrappers._concatenate:ConcatenateWorkChain"},"optimize.wrappers.create_evaluate":{description:["Wrapper workchain to combine two processes: The first process _creates_"," a result, and the second _evaluates_ that result.",""," The purpose of this workchain is to facilitate optimization of processes"," which don't natively produce an output that can be optimized, by only"," having to add the 'evaluation' part."],spec:{inputs:[{name:"create",required:!0,valid_types:"",info:"Inputs which are passed on to the create sub-process."},{name:"create_process",required:!0,valid_types:"Str",info:"The sub-process which performs the create step."},{name:"evaluate_process",required:!0,valid_types:"Str",info:"The sub-process which performs the evaluate step."},{name:"output_input_mapping",required:!0,valid_types:"Dict",info:"A mapping from output names of the create process to input names of the evaluate process. These outputs (if present) are forwarded to the evaluate process."},{name:"evaluate",required:!1,valid_types:"",info:"Inputs which are passed on to the evaluate sub-process."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"create",required:!0,valid_types:"",info:""},{name:"evaluate",required:!0,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"Workchain failed because the 'create' sub-process failed."},{status:202,message:"Workchain failed because the 'evaluate' sub-process failed."}]},class:"aiida_optimize.wrappers._create_evaluate:CreateEvaluateWorkChain"}}},commits_count:0,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"green",text:"Workflows",count:4}],pip_install_cmd:"pip install aiida-optimize",is_installable:"True"},"aiida-orca":{code_home:"https://github.com/pzarabadip/aiida-orca",development_status:"stable",documentation_url:"https://aiida-orca.readthedocs.io/",entry_point_prefix:"orca",pip_url:"git+https://github.com/pzarabadip/aiida-orca",name:"aiida-orca",package_name:"aiida_orca",hosted_on:"github.com",metadata:{author:"Pezhman Zarabadi-Poor",author_email:"pzarabadip@gmail.com",version:"0.5.1",description:"AiiDA plugin for ORCA code",classifiers:["Environment :: Plugins","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Framework :: AiiDA"]},aiida_version:">=1.0.0,<2.0.0",entry_points:{"aiida.calculations":{orca_main:"aiida_orca.calculations:OrcaCalculation",orca_asa:"aiida_orca.calculations:OrcaAsaCalculation"},"aiida.parsers":{orca_base_parser:"aiida_orca.parsers:OrcaBaseParser"},"aiida.workflows":{"orca.base":{description:["Workchain to run a orca calculation with automated error handling and restarts."],spec:{inputs:[{name:"orca",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"the results of the calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"relaxed_structure",required:!1,valid_types:"StructureData",info:"relaxed structure"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed with an unidentified unrecoverable error."},{status:301,message:"The sub process excepted."},{status:301,message:"The calculation failed with an unrecoverable error coming from aiida-orca."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_orca.workchains:OrcaBaseWorkChain"}}},commits_count:0,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/pzarabadip/aiida-orca",is_installable:"True"},"aiida-phonopy":{code_home:"https://github.com/aiida-phonopy/aiida-phonopy",documentation_url:"https://aiida-phonopy.readthedocs.io/",entry_point_prefix:"phonopy",pip_url:"aiida-phonopy",plugin_info:"https://raw.githubusercontent.com/aiida-phonopy/aiida-phonopy/master/setup.json",name:"aiida-phonopy",package_name:"aiida_phonopy",hosted_on:"github.com",metadata:{release_date:"2023-12-09",description:"The official AiiDA plugin for Phonopy",author_email:"Lorenzo Bastonero ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics"],version:"1.1.4"},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.calculations":{"phonopy.phonopy":{description:["Base `CalcJob` implementation for Phonopy post-processing."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:`Phonopy parameters (\`setting tags\`) for post processing. The following tags, along their type, are allowed: +PRIMITIVE_AXES +PRIMITIVE_AXIS +EIGENVECTORS +BAND +BAND_PATHS +BAND_POINTS +BAND_LABELS +BAND_CONNECTION +BAND_INDICES +MESH +MP +MESH_NUMBERS +MP_SHIFT +GAMMA_CENTER +WRITE_MESH +DOS +DOS_RANGE +FMIN +FMAX +FPITCH +PDOS +PROJECTION_DIRECTION +XYZ_DIRECTION +SIGMA +DEBYE_MODEL +MOMEMT +MOMENT_ORDER +TPROP +TMIN +TMAX +TSTEP +PRETEND_REAL +CUTOFF_FREQUENCY +TDISP +TDISPMAT +TDISPMAT_CIF +QPOINTS +WRITEDM +NAC_METHOD +Q_DIRECTION +GROUP_VELOCITY +GV_DELTA_Q +SYMMETRY_TOLERANCE +SYMMETRY +MESH_SYMMETRY +FC_SYMMETRY +FULL_FORCE_CONSTANTS +WRITE_FORCE_CONSTANTS +ANIME_TYPE +ANIME +MODULATION +IRREPS +SHOW_IRREPS +LITTLE_COGROUP`},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"force_constants",required:!1,valid_types:"ForceConstantsData, NoneType",info:"Force constants of the input structure."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"phonopy_data",required:!1,valid_types:"PhonopyData, NoneType",info:"The preprocess output info of a previous ForceConstantsWorkChain."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Settings for phonopy calculation."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"irreducible_representations",required:!1,valid_types:"Dict",info:"Irreducible representation output."},{name:"modulation",required:!1,valid_types:"Dict",info:"Modulation information."},{name:"output_force_constants",required:!1,valid_types:"ArrayData",info:"Calculated force constants."},{name:"output_parameters",required:!1,valid_types:"Dict",info:"Sum up info of phonopy calculation."},{name:"phonon_bands",required:!1,valid_types:"BandsData",info:"Calculated phonon band structure."},{name:"projected_phonon_dos",required:!1,valid_types:"XyData",info:"Calculated projected DOS."},{name:"qpoints",required:!1,valid_types:"BandsData",info:"Calculated qpoints."},{name:"qpoints_mesh",required:!1,valid_types:"BandsData",info:"Calculated qpoint mesh."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"thermal_displacement_matrices",required:!1,valid_types:"Dict",info:"Calculated thermal displacements matrices."},{name:"thermal_displacements",required:!1,valid_types:"Dict",info:"Calculated thermal displacements."},{name:"thermal_properties",required:!1,valid_types:"XyData",info:"Calculated thermal properties."},{name:"total_phonon_dos",required:!1,valid_types:"XyData",info:"Calculated total DOS."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The retrieved folder did not contain the required phonopy file."},{status:304,message:"The retrieved folder did not contain one or more expected output files."},{status:305,message:"No run mode has been selected."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:320,message:"The loading of yaml file got an unexpected error."},{status:321,message:"The file loading via numpy got an unexpected error."},{status:350,message:"The parser raised an unexpected exception."},{status:400,message:"The parser was not able to parse one or more files."}]},class:"aiida_phonopy.calculations.phonopy:PhonopyCalculation"}},"aiida.data":{"phonopy.force_constants":"aiida_phonopy.data.force_constants:ForceConstantsData","phonopy.phonopy":"aiida_phonopy.data.phonopy:PhonopyData","phonopy.preprocess":"aiida_phonopy.data.preprocess:PreProcessData","phonopy.raw":"aiida_phonopy.data.raw:RawData"},"aiida.parsers":{"phonopy.phonopy":"aiida_phonopy.parsers.phonopy:PhonopyParser"},"aiida.workflows":{"phonopy.phonopy":{description:["Abstract workflow for automated frozen phonons.",""," Phonopy is used to produce structures with displacements,"," while the forces are calculated with a quantum engine of choice.",""," This workchain is meant to be used as a base for other specific force calculato plugin workchains,"," or as an example on how to set a possible workchain/workflow. For this reason, the outline of"," this class is not defined, while it provides the inputs and a `setup` method, which can be used"," in a specific workflow outline. Ideally, the workflow would look like:",""," 1. Setup the preprocess data.",""," This is already provided in this class. It setups a `PreProcessData` node, from where"," supercell, primitive cell and supercells with displacements can be easily extracted using"," the methods of the nodes. This node can be taken from `self.ctx.preprocess_data`, and used"," during the outline of the workflow.",""," 2. Run supercells using the selected quantum engine/force calculator code.",""," In specific code implementations, a force calculation on supercells needs to be run."," To get these supercells, one need simply to run:",""," ```self.ctx.preprocess_data.calcfunctions.get_supercells_with_displacements()```",""," This will return a dictionary with all the supercells as StructureData to run for the phonon calculation."," The keys of this dictionary are of the type `supercell_{number}`, where `number` is an integer."," These numbers are essentials since the `phonopy` force sets is generated following these numbers,"," in order to make sure to refer to the correct displacement. Thus, it is required to keep track"," of them."," Moreover,a calculation over the pristine supercell structure should be run before hand as reference."," This structure can instead be gotten via:",""," ```self.ctx.preprocess_data.calcfunctions.get_supercell()```",""," This will return a StructureData without any label.",""," For an example of implementation, refer to aiidateam/aiida-common-worfklows.",""," * Note: some type of force calculation needs to map some variables from the unitcell to the supercell"," (and in certain case even the primitive cell), e.g. the atomic spin in VASP. Since this is code dependent,"," you will need to map these parameters before launching the force calculation of a certain supercell"," with displacement. This information can be gotten via:",""," ```self.ctx.preprocess_data.get_cells_mappings()```",""," Moreover, consider that cells in phonopy will always (re)fold the atoms in order to have positive coordinates.",""," 3. Inspect all runs and expose the forces and energies (not mandatory) outputs.",""," * Suggested: when the calculation on each supercell has finished (correctly)"," expose the output forces (and energies) in the dynamical `supercells_forces(energies)` namespace(s)."," Provide each supercell forces as an `ArrayData` with the forces stored as `forces`"," (e.g. if your code plugin stores the forces in `TrajectoryData`, extract them with a `calcfunction`)."," Expose each `ArrayData` choosing a **common prefix**, while as **suffix use"," _{number}**, with `{number}` referring to the correspective supercell label suffix (that you are supposed to"," keep track somewhere, e.g. in the label of the code calculation/workchain)."," Now you can gather all the information in one data noe, i.e. in a `PhonopyData` node."," To do so, you can simple run:",""," ```self.ctx.preprocess_data.calcfunctions.generate_phonopy_data(**self.outputs.supercells_forces)```",""," and then expose it as output in the `output_phonopy_data` namespace.",""," * Alternatively: instead of exposing the supercell forces as outputs, you can directly gather all the forces"," in a dictionary and run directly to the `generate_phonopy_data` method using this dictionary (always using"," the double *).",""," See the implementation in aiidateam/aiida-common-workflows for an example.",""," 4. (optional) Run the non-analytical constants on the primitive cell.",""," Non-analytical constants should be run for polar insulators. These require usually a linear response code"," or a finite difference approach (e.g. using the electric enthalpy). Since this is usually the most expensive"," part, you should run them on the primitive cell. To get it, use:",""," ```self.ctx.preprocess_data.calcfunctions.get_primitive_cell()```",""," If you compute also these, collect the dielectric tensor and the effectic born charges in an ArrayData,"," with the arraynames `dielectric` and `born_charges` (in Cartesian coordinates!)."," Then, gather all the information of nac and forces in a unique `PhonopyData` via:",""," ```"," self.ctx.preprocess_data.calcfunctions.generate_phonopy_data("," nac_parameters=nac_paramters,"," **self.outputs.supercells_forces"," )"," ```",""," and expose the output.",""," * Note: we require in the input for generating the full phonopy data, to give the nac in the primitive cell."," The primitive cell of phonopy will just rotate the lattice vectors, thus mantaining the Cartasian coordinate"," system. It can happen, though, that the unitcell is not the primitive cell of the system, meaning that the"," primitive cell will contain less atoms. We expect in input the nac computed on this number of atoms. If you"," want, for some reason, compute the nac on the unitcell, you will need to get the reduced nac."," To do so, you can consider using a built-in function in phonopy, namely:",""," :py:func:`phonopy.structure.symmetry.elaborate_borns_and_epsilon`"],spec:{inputs:[{name:"options",required:!0,valid_types:"",info:"Options for how to run the workflow."},{name:"displacement_generator",required:!1,valid_types:"Dict, NoneType",info:`Info for displacements generation. The following flags are allowed: + distance + is_plusminus + is_diagonal + is_trigonal + number_of_snapshots + random_seed + temperature + cutoff_frequency`},{name:"fc_options",required:!1,valid_types:"Dict, NoneType",info:`Options for force constants calculation (optional). The following flags are allowed: + calculate_full_force_constants + fc_calculator + fc_calculator_options`},{name:"is_symmetry",required:!1,valid_types:"Bool, NoneType",info:"Whether using or not the space group symmetries."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nac_parameters",required:!1,valid_types:"ArrayData, NoneType",info:"Non-analytical parameters."},{name:"preprocess_data",required:!1,valid_types:"PhonopyData, PreProcessData, NoneType",info:"The preprocess data for frozen phonon calcualtion."},{name:"primitive_matrix",required:!1,valid_types:"List, NoneType",info:"The matrix used to generate the primitive cell from the input structure in the List format. Allowed shapes are 3x1 and 3x3 lists."},{name:"structure",required:!1,valid_types:"StructureData, NoneType",info:"The structure at equilibrium volume."},{name:"supercell_matrix",required:!1,valid_types:"List, NoneType",info:"The matrix used to generate the supercell from the input structure in the List format. Allowed shapes are 3x1 and 3x3 lists."},{name:"symmetry_tolerance",required:!1,valid_types:"Float, NoneType",info:"Symmetry tolerance for space group analysis on the input structure."}],outputs:[{name:"output_phonopy_data",required:!0,valid_types:"PhonopyData",info:"The phonopy data with supercells displacements, forces and (optionally)nac parameters to use in the post-processing calculation."},{name:"supercells_forces",required:!0,valid_types:"ArrayData",info:"The forces acting on the atoms of each supercell."},{name:"output_force_constants",required:!1,valid_types:"ForceConstantsData",info:"The matrix of force constants computed with finite displacements."},{name:"supercells",required:!1,valid_types:"StructureData",info:"The supercells with displacements."},{name:"supercells_energies",required:!1,valid_types:"Float",info:"The total energy of each supercell."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_phonopy.workflows.phonopy:PhonopyWorkChain"}}},commits_count:8,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:4},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-phonopy",is_installable:"True"},"aiida-phtools":{code_home:"https://github.com/ltalirz/aiida-phtools",entry_point_prefix:"phtools",pip_url:"aiida-phtools",plugin_info:"https://raw.github.com/ltalirz/aiida-phtools/master/setup.json",name:"aiida-phtools",package_name:"aiida_phtools",hosted_on:"github.com",metadata:{release_date:"2018-06-21",description:"AiiDA plugin for persistence homology tools, used to analyze nanoporous materials.",author:"Leopold Talirz",author_email:"leopold.talirz@gmail.com",license:"MIT",home_page:"https://github.com/ltalirz/aiida-phtools",classifiers:["Programming Language :: Python"],version:"0.1.0a1"},aiida_version:"*",entry_points:{"aiida.calculations":{"phtools.dmatrix":"aiida_phtools.calculations.distance_matrix:DistanceMatrixCalculation","phtools.surface":"aiida_phtools.calculations.pore_surface:PoreSurfaceCalculation"},"aiida.data":{"phtools.surface":"aiida_phtools.data.pore_surface:PoreSurfaceParameters"},"aiida.parsers":{"phtools.dmatrix":"aiida_phtools.parsers.distance_matrix:DistanceMatrixParser","phtools.surface":"aiida_phtools.parsers.pore_surface:PoreSurfaceParser"}},commits_count:0,development_status:"planning",errors:[],warnings:["W003: Missing classifier 'Framework :: AiiDA'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:1}],pip_install_cmd:"pip install --pre aiida-phtools"},"aiida-plumed":{code_home:"https://github.com/ConradJohnston/aiida-plumed",entry_point_prefix:"plumed",pip_url:"aiida-plumed",plugin_info:"https://raw.github.com/ConradJohnston/aiida-plumed/AiiDA-v1.0-compatibility/setup.json",name:"aiida-plumed",package_name:"aiida_plumed",hosted_on:"github.com",metadata:{release_date:"2019-09-16",description:"AiiDA plugin providing support for Plumed2",author:"Conrad Johnston",author_email:"conrad.s.johnston@googlemail.com",license:"MIT",home_page:"https://github.com/ConradJohnston/aiida-plumed",classifiers:["Development Status :: 2 - Pre-Alpha","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.1.0a0"},aiida_version:">=1.0.0b3,<2.0.0",entry_points:{"aiida.calculations":{plumed:"aiida_plumed.calculations:DiffCalculation"},"aiida.cmdline.data":{plumed:"aiida_plumed.cli:data_cli"},"aiida.data":{plumed:"aiida_plumed.data:DiffParameters"},"aiida.parsers":{plumed:"aiida_plumed.parsers:DiffParser"}},commits_count:0,development_status:"pre-alpha",errors:[`E001: Failed to install plugin aiida-plumed
Collecting aiida-plumed
+  Downloading aiida-plumed-0.1.0a0.tar.gz (11 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-plumed to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-plumed==0.1.0a0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-plumed 0.1.0a0 depends on aiida-core<2.0.0 and >=1.0.0b3
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install --pre aiida-plumed",is_installable:"False"},"aiida-porousmaterials":{code_home:"https://github.com/pzarabadip/aiida-porousmaterials",development_status:"stable",entry_point_prefix:"porousmaterials",pip_url:"aiida-porousmaterials",name:"aiida-porousmaterials",package_name:"aiida_porousmaterials",hosted_on:"github.com",metadata:{release_date:"2020-03-05",description:"AiiDA plugin for PorousMaterials code",author:"Pezhman Zarabadi-Poor",author_email:"pzarabadip@gmail.com",license:"MIT",home_page:"https://github.com/pzarabadip/aiida-porousmaterials",classifiers:["Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8"],version:"1.0.0a3"},aiida_version:null,entry_points:{"aiida.calculations":{porousmaterials:"aiida_porousmaterials.calculations:PorousMaterialsCalculation"},"aiida.parsers":{porousmaterials:"aiida_porousmaterials.parser:PorousMaterialsParser"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-porousmaterials
Collecting aiida-porousmaterials
+  Downloading aiida-porousmaterials-1.0.0a3.tar.gz (5.7 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-porousmaterials to determine which version is compatible with other requirements. This could take a while.
+  Downloading aiida_porousmaterials-1.0.0a2-py3-none-any.whl.metadata (2.6 kB)
+  Downloading aiida_porousmaterials-1.0.0a1-py3-none-any.whl.metadata (2.6 kB)
+ERROR: Cannot install aiida-porousmaterials==1.0.0a1, aiida-porousmaterials==1.0.0a2 and aiida-porousmaterials==1.0.0a3 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-porousmaterials 1.0.0a3 depends on aiida_core<2.0.0 and >=1.0.0
+    aiida-porousmaterials 1.0.0a2 depends on aiida-core<2.0.0 and >=1.0.0
+    aiida-porousmaterials 1.0.0a1 depends on aiida-core<2.0.0 and >=1.0.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install --pre aiida-porousmaterials",is_installable:"False"},"aiida-pseudo":{code_home:"https://github.com/aiidateam/aiida-pseudo",entry_point_prefix:"pseudo",pip_url:"aiida-pseudo",plugin_info:"https://raw.github.com/aiidateam/aiida-pseudo/master/setup.cfg",name:"aiida-pseudo",package_name:"aiida_pseudo",hosted_on:"github.com",metadata:{release_date:"2024-11-04",description:"AiiDA plugin that simplifies working with pseudo potentials.",author_email:'"Sebastiaan P. Huber" ',classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9"],version:"1.6.0"},aiida_version:">=2.6,<3.0",entry_points:{"aiida.data":{pseudo:"aiida_pseudo.data.pseudo.pseudo:PseudoPotentialData","pseudo.jthxml":"aiida_pseudo.data.pseudo.jthxml:JthXmlData","pseudo.psf":"aiida_pseudo.data.pseudo.psf:PsfData","pseudo.psml":"aiida_pseudo.data.pseudo.psml:PsmlData","pseudo.psp8":"aiida_pseudo.data.pseudo.psp8:Psp8Data","pseudo.upf":"aiida_pseudo.data.pseudo.upf:UpfData","pseudo.vps":"aiida_pseudo.data.pseudo.vps:VpsData"},"aiida.groups":{"pseudo.family":"aiida_pseudo.groups.family.pseudo:PseudoPotentialFamily","pseudo.family.cutoffs":"aiida_pseudo.groups.family.cutoffs:CutoffsPseudoPotentialFamily","pseudo.family.pseudo_dojo":"aiida_pseudo.groups.family.pseudo_dojo:PseudoDojoFamily","pseudo.family.sssp":"aiida_pseudo.groups.family.sssp:SsspFamily"},console_scripts:{"aiida-pseudo":"aiida_pseudo.cli:cmd_root"}},commits_count:18,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"red",text:"Data",count:7},{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Groups)",count:4}],pip_install_cmd:"pip install aiida-pseudo",is_installable:"True"},"aiida-psi4":{code_home:"https://github.com/ltalirz/aiida-psi4/tree/master",development_status:"beta",entry_point_prefix:"psi4",pip_url:"git+https://github.com/ltalirz/aiida-psi4",name:"aiida-psi4",package_name:"aiida_psi4",hosted_on:"github.com",metadata:{author:"Leopold Talirz",author_email:"leopold.talirz@gmail.com",version:"0.1.0a0",description:"AiiDA plugin for the Psi4 Quantum Chemistry package.",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=1.6.4,<2.0.0",entry_points:{"aiida.data":{"psi4.atomic_input":"aiida_psi4.data:AtomicInput"},"aiida.calculations":{psi4:"aiida_psi4.calculations:Psi4Calculation"},"aiida.parsers":{psi4:"aiida_psi4.parsers:QCSchemaParser"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-psi4
Collecting git+https://github.com/ltalirz/aiida-psi4
+  Cloning https://github.com/ltalirz/aiida-psi4 to /tmp/pip-req-build-cfuplcbg
+  Running command git clone --filter=blob:none --quiet https://github.com/ltalirz/aiida-psi4 /tmp/pip-req-build-cfuplcbg
+  Resolved https://github.com/ltalirz/aiida-psi4 to commit 637e6b0b29e724a158014269d55d9091c6af48c7
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Installing backend dependencies: started
+  Installing backend dependencies: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-psi4 to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-psi4==0.1.0a0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-psi4 0.1.0a0 depends on aiida-core<2.0.0 and >=1.6.4
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1}],pip_install_cmd:"pip install git+https://github.com/ltalirz/aiida-psi4",is_installable:"False"},"aiida-pyscf":{code_home:"https://github.com/microsoft/aiida-pyscf",entry_point_prefix:"pyscf",pip_url:"aiida-pyscf",plugin_info:"https://raw.githubusercontent.com/microsoft/aiida-pyscf/main/pyproject.toml",name:"aiida-pyscf",package_name:"aiida_pyscf",hosted_on:"github.com",metadata:{release_date:"2024-01-03",description:"AiiDA plugin for the Python-based Simulations of Chemistry Framework (PySCF).",author_email:'"Sebastiaan P. Huber" , Adam Grofe ',classifiers:["Development Status :: 3 - Alpha","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering"],version:"0.5.1"},aiida_version:">=2.5,<3.0",entry_points:{"aiida.calculations":{"pyscf.base":{description:["``CalcJob`` plugin for PySCF."],spec:{inputs:[{name:"code",required:!0,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure with molecular structure definition."},{name:"checkpoint",required:!1,valid_types:"SinglefileData, NoneType",info:"Checkpoint of a previously completed calculation that failed to converge."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Input parameters used to render the PySCF script template."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"checkpoint",required:!1,valid_types:"SinglefileData",info:"The checkpoint file in case the calculation did not converge. Can be used as an input for a restart."},{name:"cubegen",required:!1,valid_types:"",info:""},{name:"fcidump",required:!1,valid_types:"SinglefileData",info:"Computed fcidump files."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The computed Hessian."},{name:"model",required:!1,valid_types:"PickledData",info:"The model in serialized form. Can be deserialized and used without having to run the kernel again."},{name:"parameters",required:!1,valid_types:"Dict",info:"Various computed properties parsed from the `FILENAME_RESULTS` output file."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"structure",required:!1,valid_types:"StructureData",info:"The optimized structure if the input parameters contained the `optimizer` key."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The geometry optimization trajectory if the input parameters contained the `optimizer` key."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The stdout output file was not retrieved."},{status:303,message:"The results JSON file was not retrieved."},{status:410,message:"The electronic minimization cycle did not reach self-consistency."},{status:500,message:"The ionic minimization cycle did not converge for the given thresholds."}]},class:"aiida_pyscf.calculations.base:PyscfCalculation"}},"aiida.parsers":{"pyscf.base":"aiida_pyscf.parsers.base:PyscfParser"},"aiida.workflows":{"pyscf.base":{description:["Workchain to run a pyscf calculation with automated error handling and restarts."],spec:{inputs:[{name:"pyscf",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"checkpoint",required:!1,valid_types:"SinglefileData",info:"The checkpoint file in case the calculation did not converge. Can be used as an input for a restart."},{name:"cubegen",required:!1,valid_types:"",info:""},{name:"fcidump",required:!1,valid_types:"SinglefileData",info:"Computed fcidump files."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The computed Hessian."},{name:"model",required:!1,valid_types:"PickledData",info:"The model in serialized form. Can be deserialized and used without having to run the kernel again."},{name:"parameters",required:!1,valid_types:"Dict",info:"Various computed properties parsed from the `FILENAME_RESULTS` output file."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"structure",required:!1,valid_types:"StructureData",info:"The optimized structure if the input parameters contained the `optimizer` key."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The geometry optimization trajectory if the input parameters contained the `optimizer` key."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed with an unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:310,message:"The calculation failed and did not retrieve a checkpoint file from which can be restarted."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_pyscf.workflows.base:PyscfBaseWorkChain"}}},commits_count:13,development_status:"alpha",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-pyscf",is_installable:"True"},"aiida-python":{entry_point_prefix:"aiida-python",development_status:"beta",code_home:"https://github.com/addman2/aiida-python",pip_url:"aiida-python",documentation_url:"https://aiida-python.readthedocs.io",name:"aiida-python",package_name:"aiida_python",hosted_on:"github.com",metadata:{release_date:"2023-11-03",description:"AiiDA Python",author:"Ot(t)o Kohulák",author_email:"otto.kohulak@dontemail.me",license:"MIT",home_page:"https://todo.nothing/",classifiers:["Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.0.6"},aiida_version:">=2.1.0",entry_points:{"aiida.calculations":{"python.calc":"aiida_python.calc:CalcJobPython"},"aiida.parsers":{"python.parser":"aiida_python.parser:ParserPython"},"aiida_python.serializers":{int:"aiida_python.serializers:SerializerInt",float:"aiida_python.serializers:SerializerFloat",str:"aiida_python.serializers:SerializerStr",list:"aiida_python.serializers:SerializerList",arraydata:"aiida_python.serializers:SerializerArrayData"},"aiida.data":{}},commits_count:4,errors:[],warnings:["W019: No bdist_wheel available for PyPI release.","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W009: Prefix 'aiida-python' does not follow naming convention.","W010: Entry point 'python.calc' does not start with prefix 'aiida-python.'","W010: Entry point 'python.parser' does not start with prefix 'aiida-python.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"orange",text:"Other (Aiida python serializers)",count:5}],pip_install_cmd:"pip install aiida-python",is_installable:"True"},"aiida-qeq":{code_home:"https://github.com/ltalirz/aiida-qeq",development_status:"stable",entry_point_prefix:"qeq",pip_url:"aiida-qeq",plugin_info:"https://raw.githubusercontent.com/ltalirz/aiida-qeq/master/setup.json",name:"aiida-qeq",package_name:"aiida_qeq",hosted_on:"github.com",metadata:{release_date:"2018-11-21",description:"AiiDA plugin for computing electronic charges on atoms using equilibration-type models (QEq, EQEq, ...).",author:"Leopold Talirz, Daniele Ongari",author_email:"leopold.talirz@gmail.com",license:"MIT",home_page:"https://github.com/ltalirz/aiida-qeq",classifiers:["Programming Language :: Python"],version:"0.1.0"},aiida_version:">=0.12.2,<1.0.0",entry_points:{"aiida.calculations":{"qeq.eqeq":"aiida_qeq.calculations.eqeq:EQeqCalculation","qeq.qeq":"aiida_qeq.calculations.qeq:QeqCalculation"},"aiida.data":{"qeq.eqeq":"aiida_qeq.data.eqeq:EQeqParameters","qeq.qeq":"aiida_qeq.data.qeq:QeqParameters"},"aiida.parsers":{"qeq.eqeq":"aiida_qeq.parsers.eqeq:EQeqParser","qeq.qeq":"aiida_qeq.parsers.qeq:QeqParser"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-qeq
Collecting aiida-qeq
+  Downloading aiida-qeq-0.1.0.tar.gz (10 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-qeq to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-qeq==0.1.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-qeq 0.1.0 depends on aiida-core<1.0.0 and >=0.12.2
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:2}],pip_install_cmd:"pip install aiida-qeq",is_installable:"False"},"aiida-qp2":{code_home:"https://github.com/TREX-CoE/aiida-qp2",entry_point_prefix:"qp2",pip_url:"aiida-qp2",documentation_url:"https://trex-coe.github.io/aiida-qp2/index.html",name:"aiida-qp2",package_name:"aiida_qp2",hosted_on:"github.com",metadata:{release_date:"2022-02-26",description:"AiiDA plugin for the Quantum Package 2.0",author:"Evgeny Posenitskiy",author_email:"posenitskiy@irsamc.ups-tlse.fr",license:"MIT",home_page:"https://github.com/TREX-CoE/aiida-qp2",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Programming Language :: Python"],version:"0.2.0"},aiida_version:null,entry_points:{console_scripts:{aqp:"aiida_qp2.cli:cli_root"},"aiida.cmdline.data":{"qp2.cli":"aiida_qp2.cli:cli_root"},"aiida.calculations":{qp2:"aiida_qp2.calculations:QP2Calculation","qp2.create":"aiida_qp2.create.calculation:QP2CreateCalculation","qp2.run":"aiida_qp2.run.calculation:QP2RunCalculation","qp2.qmcchemrun":"aiida_qp2.run.qmcchem_calculation:QP2QmcchemRunCalculation"},"aiida.parsers":{qp2:"aiida_qp2.parsers:QP2Parser","qp2.create":"aiida_qp2.create.parser:QP2CreateParser","qp2.run":"aiida_qp2.run.parser:QP2RunParser","qp2.qmcchemrun":"aiida_qp2.run.qmcchem_parser:QP2QmcchemRunParser"}},commits_count:55,development_status:"beta",errors:[`E001: Failed to install plugin aiida-qp2
Collecting aiida-qp2
+  Downloading aiida-qp2-0.2.0.tar.gz (9.8 kB)
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Installing backend dependencies: started
+  Installing backend dependencies: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-qp2 to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-qp2==0.2.0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-qp2 0.2.0 depends on aiida-core<2.0.0 and >=1.1.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:4},{colorclass:"brown",text:"Parsers",count:4},{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Data commands)",count:1}],pip_install_cmd:"pip install aiida-qp2",is_installable:"False"},"aiida-quantumespresso":{code_home:"https://github.com/aiidateam/aiida-quantumespresso",documentation_url:"https://aiida-quantumespresso.readthedocs.io/",entry_point_prefix:"quantumespresso",pip_url:"aiida-quantumespresso",plugin_info:"https://raw.github.com/aiidateam/aiida-quantumespresso/master/setup.json",name:"aiida-quantumespresso",package_name:"aiida_quantumespresso",hosted_on:"github.com",metadata:{release_date:"2024-04-23",description:"The official AiiDA plugin for Quantum ESPRESSO",author_email:"The AiiDA team ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"4.6.0"},aiida_version:">=2.3,<3.0",entry_points:{"aiida.calculations":{"quantumespresso.cp":{description:["`CalcJob` implementation for the cp.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"The input parameters that are to be used to construct the input file."},{name:"pseudos",required:!0,valid_types:"UpfData, UpfData",info:"A mapping of `UpfData` nodes onto the kind name to which they should apply."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parallelization",required:!1,valid_types:"Dict, NoneType",info:`Parallelization options. The following flags are allowed: +`},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"An optional working directory of a previously completed calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job and the parsing are performed."},{name:"vdw_table",required:!1,valid_types:"SinglefileData, NoneType",info:"Optional van der Waals table contained in a `SinglefileData`."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"output_trajectory",required:!0,valid_types:"TrajectoryData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The required XML file is not present in the retrieved folder."},{status:304,message:"The retrieved folder contains multiple XML files."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:320,message:"The required XML file could not be read."},{status:330,message:"The required POS file could not be read."},{status:340,message:"The required trajectory data could not be read."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."}]},class:"aiida_quantumespresso.calculations.cp:CpCalculation"},"quantumespresso.create_kpoints_from_distance":{description:["Generate a uniformly spaced kpoint mesh for a given structure.",""," The spacing between kpoints in reciprocal space is guaranteed to be at least the defined distance.",""," :param structure: the StructureData to which the mesh should apply"," :param distance: a Float with the desired distance between kpoints in reciprocal space"," :param force_parity: a Bool to specify whether the generated mesh should maintain parity"," :returns: a KpointsData with the generated mesh"],spec:{inputs:[{name:"distance",required:!0,valid_types:"Data",info:"a Float with the desired distance between kpoints in reciprocal space"},{name:"force_parity",required:!0,valid_types:"Data",info:"a Bool to specify whether the generated mesh should maintain parity"},{name:"structure",required:!0,valid_types:"Data",info:"the StructureData to which the mesh should apply"},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_quantumespresso.calculations.functions.create_kpoints_from_distance:create_kpoints_from_distance"},"quantumespresso.create_magnetic_configuration":{description:["Create a new magnetic configuration from the given structure based on a list of magnetic moments per site.",""," To create the new list of kinds, the algorithm loops over all the elements in the structure and makes a list of the"," sites with that element and their corresponding magnetic moment. Next, it splits this list in three lists:",""," * Zero magnetic moments: Any site that has an absolute magnetic moment lower than ``ztol``"," * Positive magnetic moments"," * Negative magnetic moments",""," The algorithm then sorts the positive and negative lists from large to small absolute value, and loops over each of"," list. New magnetic kinds will be created when the absolute difference between the magnetic moment of the current"," kind and the site exceeds ``atol``.",""," The positive and negative magnetic moments are handled separately to avoid assigning two sites with opposite signs"," in their magnetic moment to the same kind and make sure that each kind has the correct magnetic moment, i.e. the"," largest magnetic moment in absolute value of the sites corresponding to that kind.",""," .. important:: the function currently does not support alloys.",""," :param structure: a `StructureData` instance."," :param magnetic_moment_per_site: list of magnetic moments for each site in the structure."," :param atol: the absolute tolerance on determining if two sites have the same magnetic moment."," :param ztol: threshold for considering a kind to have non-zero magnetic moment."],spec:{inputs:[{name:"magnetic_moment_per_site",required:!0,valid_types:"Data",info:"list of magnetic moments for each site in the structure."},{name:"structure",required:!0,valid_types:"Data",info:"a `StructureData` instance."},{name:"atol",required:!1,valid_types:"Data",info:"the absolute tolerance on determining if two sites have the same magnetic moment."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"ztol",required:!1,valid_types:"Data",info:"threshold for considering a kind to have non-zero magnetic moment."}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_quantumespresso.calculations.functions.create_magnetic_configuration:create_magnetic_configuration"},"quantumespresso.dos":{description:["`CalcJob` implementation for the dos.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parent_folder",required:!0,valid_types:"RemoteData, FolderData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"output_dos",required:!0,valid_types:"XyData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:330,message:"The dos file could not be read from the retrieved folder."}]},class:"aiida_quantumespresso.calculations.dos:DosCalculation"},"quantumespresso.epw":{description:["`CalcJob` implementation for the epw.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"kfpoints",required:!0,valid_types:"KpointsData",info:"fine kpoint mesh"},{name:"kpoints",required:!0,valid_types:"KpointsData",info:"coarse kpoint mesh"},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"parent_folder_nscf",required:!0,valid_types:"RemoteData",info:"the folder of a completed nscf `PwCalculation`"},{name:"parent_folder_ph",required:!0,valid_types:"RemoteData",info:"the folder of a completed `PhCalculation`"},{name:"qfpoints",required:!0,valid_types:"KpointsData",info:"fine qpoint mesh"},{name:"qpoints",required:!0,valid_types:"KpointsData",info:"coarse qpoint mesh"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_quantumespresso.calculations.epw:EpwCalculation"},"quantumespresso.matdyn":{description:["`CalcJob` implementation for the matdyn.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"force_constants",required:!0,valid_types:"ForceConstantsData",info:""},{name:"kpoints",required:!0,valid_types:"KpointsData",info:"Kpoints on which to calculate the phonon frequencies."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"parent_folder",required:!1,valid_types:"RemoteData, FolderData, SinglefileData, NoneType",info:"Use a local or remote folder as parent folder (for restarts and similar)"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"output_phonon_bands",required:!0,valid_types:"BandsData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:330,message:"The output frequencies file could not be read from the retrieved folder."},{status:410,message:"Number of kpoints not found in the output data"},{status:411,message:"Number of kpoints in the inputs is not commensurate with those in the output"}]},class:"aiida_quantumespresso.calculations.matdyn:MatdynCalculation"},"quantumespresso.merge_ph_outputs":{description:["Calcfunction to merge outputs from multiple `ph.x` calculations with different q-points."],spec:{inputs:[{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_quantumespresso.calculations.functions.merge_ph_outputs:merge_ph_outputs"},"quantumespresso.namelists":{description:["`CalcJob` implementation to serve as base class for simple post-processing tools of Quantum ESPRESSO."],spec:{inputs:[{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"parent_folder",required:!1,valid_types:"RemoteData, FolderData, SinglefileData, NoneType",info:"Use a local or remote folder as parent folder (for restarts and similar)"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."}]},class:"aiida_quantumespresso.calculations.namelists:NamelistsCalculation"},"quantumespresso.neb":{description:["Nudged Elastic Band code (neb.x) of Quantum ESPRESSO distribution."],spec:{inputs:[{name:"first_structure",required:!0,valid_types:"StructureData",info:"Initial structure"},{name:"last_structure",required:!0,valid_types:"StructureData",info:"Final structure"},{name:"parameters",required:!0,valid_types:"Dict",info:"NEB-specific input parameters"},{name:"pw",required:!0,valid_types:"Data",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"An optional working directory of a previously completed calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job and the parsing are performed."}],outputs:[{name:"output_mep",required:!0,valid_types:"ArrayData",info:"The original and interpolated energy profiles along the minimum-energy path (mep)"},{name:"output_parameters",required:!0,valid_types:"Dict",info:"The output parameters dictionary of the NEB calculation"},{name:"output_trajectory",required:!0,valid_types:"TrajectoryData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"iteration_array",required:!1,valid_types:"ArrayData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:303,message:"The required XML file is not present in the retrieved folder."},{status:320,message:"The XML output file could not be read."},{status:321,message:"The XML output file could not be parsed."},{status:322,message:"The XML output file has an unsupported format."},{status:350,message:"The parser raised an unexpected exception: {exception}"}]},class:"aiida_quantumespresso.calculations.neb:NebCalculation"},"quantumespresso.open_grid":{description:["``CalcJob`` implementation for the ``open_grid.x`` code of Quantum ESPRESSO."],spec:{inputs:[{name:"parent_folder",required:!0,valid_types:"RemoteData, FolderData",info:"The output folder of a completed `PwCalculation` on an irreducible Brillouin zone"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The explicit list of kpoints of the unfolded kmesh"},{name:"kpoints_mesh",required:!0,valid_types:"KpointsData",info:"The dimensions of the unfolded kmesh"},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"The retrieved folder data node could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:312,message:"Found rotation or fractional translation not compatible with FFT grid."},{status:350,message:"Mismatch between kmesh dimensions and number of kpoints."}]},class:"aiida_quantumespresso.calculations.open_grid:OpenGridCalculation"},"quantumespresso.ph":{description:["`CalcJob` implementation for the ph.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"parent_folder",required:!0,valid_types:"RemoteData",info:"the folder of a completed `PwCalculation`"},{name:"qpoints",required:!0,valid_types:"KpointsData",info:"qpoint mesh"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:305,message:"Both the stdout and XML output files could not be read or parsed."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:350,message:"The parser raised an unexpected exception: {exception}"},{status:360,message:"The FFT grid is incompatible with the detected symmetries. Try using the lattice-specific `ibrav` != 0 in the parent `pw.x` calculation."},{status:361,message:"The representation found seems to be wrong according to the detected symmetries. Try using the lattice-specific `ibrav` != 0 in the parent `pw.x` calculation."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:410,message:"The minimization cycle did not reach self-consistency."},{status:462,message:"The code failed during the cholesky factorization."}]},class:"aiida_quantumespresso.calculations.ph:PhCalculation"},"quantumespresso.pp":{description:["`CalcJob` implementation for the pp.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Use a node that specifies the input parameters for the namelists"},{name:"parent_folder",required:!0,valid_types:"RemoteData, FolderData",info:"Output folder of a completed `PwCalculation`"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job is performed."}],outputs:[{name:"output_data",required:!0,valid_types:"ArrayData",info:""},{name:"output_data_multiple",required:!0,valid_types:"ArrayData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The parent folder did not contain the required XML output file."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete."},{status:330,message:"The formatted data output file `{filename}` was not present in the retrieved (temporary) folder."},{status:331,message:"The formatted data output file `{filename}` could not be read."},{status:332,message:"The data file format is not supported by the parser"},{status:333,message:"The formatted data output file `{filename}` could not be parsed: {exception}"},{status:340,message:"The calculation stopped prematurely because it ran out of walltime but the job was killed by the scheduler before the files were safely written to disk for a potential restart."},{status:350,message:"The parser raised an unexpected exception: {exception}"}]},class:"aiida_quantumespresso.calculations.pp:PpCalculation"},"quantumespresso.projwfc":{description:["`CalcJob` implementation for the projwfc.x code of Quantum ESPRESSO.",""," Projwfc.x code of the Quantum ESPRESSO distribution, handles the the computation of projections of bloch"," wavefunctions onto atomic orbitals.",""," . For more information, refer to http://www.quantum-espresso.org/"],spec:{inputs:[{name:"parent_folder",required:!0,valid_types:"RemoteData, FolderData",info:"The output folder of a pw.x calculation"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"Dos",required:!0,valid_types:"XyData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:""},{name:"bands_down",required:!1,valid_types:"BandsData",info:""},{name:"bands_up",required:!1,valid_types:"BandsData",info:""},{name:"projections",required:!1,valid_types:"ProjectionData",info:""},{name:"projections_down",required:!1,valid_types:"ProjectionData",info:""},{name:"projections_up",required:!1,valid_types:"ProjectionData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The retrieved folder did not contain the required XML file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:320,message:"The XML output file could not be read."},{status:321,message:"The XML output file could not be parsed."},{status:322,message:"The XML output file has an unsupported format."},{status:330,message:"The pdos_tot file could not be read from the retrieved folder."},{status:340,message:"An exception was raised parsing bands and projections."}]},class:"aiida_quantumespresso.calculations.projwfc:ProjwfcCalculation"},"quantumespresso.pw":{description:["`CalcJob` implementation for the pw.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"kpoint mesh or kpoint path"},{name:"parameters",required:!0,valid_types:"Dict",info:"The input parameters that are to be used to construct the input file."},{name:"pseudos",required:!0,valid_types:"UpfData, UpfData",info:"A mapping of `UpfData` nodes onto the kind name to which they should apply."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"hubbard_file",required:!1,valid_types:"SinglefileData, NoneType",info:"SinglefileData node containing the output Hubbard parameters from a HpCalculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parallelization",required:!1,valid_types:"Dict, NoneType",info:`Parallelization options. The following flags are allowed: +npool : The number of 'pools', each taking care of a group of k-points. +nband : The number of 'band groups', each taking care of a group of Kohn-Sham orbitals. +ntg : The number of 'task groups' across which the FFT planes are distributed. +ndiag : The number of 'linear algebra groups' used when parallelizing the subspace diagonalization / iterative orthonormalization. By default, no parameter is passed to Quantum ESPRESSO, meaning it will use its default.`},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"An optional working directory of a previously completed calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job and the parsing are performed."},{name:"vdw_table",required:!1,valid_types:"SinglefileData, NoneType",info:"Optional van der Waals table contained in a `SinglefileData`."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The `output_parameters` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_atomic_occupations",required:!1,valid_types:"Dict",info:""},{name:"output_band",required:!1,valid_types:"BandsData",info:"The `output_band` output node of the successful calculation if present."},{name:"output_kpoints",required:!1,valid_types:"KpointsData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The `output_structure` output node of the successful calculation if present."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The retrieved folder did not contain the required XML file."},{status:304,message:"The retrieved folder contained multiple XML files."},{status:305,message:"Both the stdout and XML output files could not be read or parsed."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:320,message:"The XML output file could not be read."},{status:321,message:"The XML output file could not be parsed."},{status:322,message:"The XML output file has an unsupported format."},{status:340,message:"The calculation stopped prematurely because it ran out of walltime but the job was killed by the scheduler before the files were safely written to disk for a potential restart."},{status:350,message:"The parser raised an unexpected exception: {exception}"},{status:360,message:"The code failed in finding a valid reciprocal lattice vector."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:410,message:"The electronic minimization cycle did not reach self-consistency."},{status:461,message:"The code failed with negative dexx in the exchange calculation."},{status:462,message:"The code failed during the cholesky factorization."},{status:463,message:"Too many bands failed to converge during the diagonalization."},{status:464,message:"The S matrix was found to be not positive definite."},{status:465,message:"The `zhegvd` failed in the PPCG diagonalization."},{status:466,message:"The `[Q, R] = qr(X, 0)` failed in the PPCG diagonalization."},{status:467,message:"The eigenvector failed to converge."},{status:468,message:"The factorization in the Broyden routine failed."},{status:481,message:'The k-point parallelization "npools" is too high, some nodes have no k-points.'},{status:500,message:"The ionic minimization cycle did not converge for the given thresholds."},{status:501,message:"Then ionic minimization cycle converged but the thresholds are exceeded in the final SCF."},{status:502,message:"The ionic minimization cycle did not converge after the maximum number of steps."},{status:503,message:"The ionic minimization cycle did not finish because the calculation was interrupted but a partial trajectory and output structure was successfully parsed which can be used for a restart."},{status:510,message:"The electronic minimization cycle failed during an ionic minimization cycle."},{status:511,message:"The ionic minimization cycle converged, but electronic convergence was not reached in the final SCF."},{status:520,message:"The ionic minimization cycle terminated prematurely because of two consecutive failures in the BFGS algorithm."},{status:521,message:"The ionic minimization cycle terminated prematurely because of two consecutive failures in the BFGS algorithm and electronic convergence failed in the final SCF."},{status:531,message:"The electronic minimization cycle did not reach self-consistency."},{status:541,message:"The variable cell optimization broke the symmetry of the k-points."},{status:542,message:"The cell relaxation caused a significant volume contraction and there is not enough space allocated for radial FFT."},{status:710,message:"The electronic minimization cycle did not reach self-consistency, but `scf_must_converge` is `False` and/or `electron_maxstep` is 0."}]},class:"aiida_quantumespresso.calculations.pw:PwCalculation"},"quantumespresso.pw2gw":{description:["`CalcJob` implementation for the pw2gw.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parent_folder",required:!0,valid_types:"RemoteData",info:"Output folder of a completed `PwCalculation`"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"eps",required:!0,valid_types:"ArrayData",info:"The `eps` output node containing 5 arrays `energy`, `epsX`, `epsY`, `epsZ`, `epsTOT`"},{name:"output_parameters",required:!0,valid_types:"Dict",info:"The `output_parameters` output node of the successful calculation.`"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:305,message:"The eps*.dat output files could not be read or parsed."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:330,message:"The eps*.dat output files do not have the expected shape (N, 2)."},{status:331,message:"The eps*.dat output files contains different values of energies."},{status:350,message:"The parser raised an unexpected exception: {exception}"}]},class:"aiida_quantumespresso.calculations.pw2gw:Pw2gwCalculation"},"quantumespresso.pw2wannier90":{description:["`CalcJob` implementation for the pw2wannier.x code of Quantum ESPRESSO.",""," For more information, refer to http://www.quantum-espresso.org/ and http://www.wannier.org/"],spec:{inputs:[{name:"nnkp_file",required:!0,valid_types:"SinglefileData",info:"A SinglefileData containing the .nnkp file generated by wannier90.x -pp"},{name:"parent_folder",required:!0,valid_types:"RemoteData, FolderData",info:"The output folder of a pw.x calculation"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:340,message:"Encountered a generic error message"},{status:350,message:"The parser raised an unexpected exception: {exception}"}]},class:"aiida_quantumespresso.calculations.pw2wannier90:Pw2wannier90Calculation"},"quantumespresso.pwimmigrant":{description:["Create a PwCalculation object that can be used to import old jobs.",""," This is a sublass of aiida_quantumespresso.calculations.PwCalculation"," with slight modifications to some of the class variables and additional"," methods that",""," a. parse the job's input file to create the calculation's input"," nodes that would exist if the calculation were submitted using AiiDa,"," b. bypass the functions of the daemon, and prepare the node's attributes"," such that all the processes (copying of the files to the repository,"," results parsing, ect.) can be performed",""," .. note:: The keyword arguments of PwCalculation are also available.",""," :param remote_workdir: Absolute path to the directory where the job was run."," The transport of the computer you link ask input to the calculation is"," the transport that will be used to retrieve the calculation's files."," Therefore, ``remote_workdir`` should be the absolute path to the job's"," directory on that computer."," :type remote_workdir: str",""," :param input_file_name: The file name of the job's input file."," :type input_file_name: str",""," :param output_file_name: The file name of the job's output file (i.e. the"," file containing the stdout of QE)."," :type output_file_name: str"],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"kpoint mesh or kpoint path"},{name:"parameters",required:!0,valid_types:"Dict",info:"The input parameters that are to be used to construct the input file."},{name:"pseudos",required:!0,valid_types:"UpfData, UpfData",info:"A mapping of `UpfData` nodes onto the kind name to which they should apply."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"hubbard_file",required:!1,valid_types:"SinglefileData, NoneType",info:"SinglefileData node containing the output Hubbard parameters from a HpCalculation"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parallelization",required:!1,valid_types:"Dict, NoneType",info:`Parallelization options. The following flags are allowed: +npool : The number of 'pools', each taking care of a group of k-points. +nband : The number of 'band groups', each taking care of a group of Kohn-Sham orbitals. +ntg : The number of 'task groups' across which the FFT planes are distributed. +ndiag : The number of 'linear algebra groups' used when parallelizing the subspace diagonalization / iterative orthonormalization. By default, no parameter is passed to Quantum ESPRESSO, meaning it will use its default.`},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"An optional working directory of a previously completed calculation to restart from."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional parameters to affect the way the calculation job and the parsing are performed."},{name:"vdw_table",required:!1,valid_types:"SinglefileData, NoneType",info:"Optional van der Waals table contained in a `SinglefileData`."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The `output_parameters` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_atomic_occupations",required:!1,valid_types:"Dict",info:""},{name:"output_band",required:!1,valid_types:"BandsData",info:"The `output_band` output node of the successful calculation if present."},{name:"output_kpoints",required:!1,valid_types:"KpointsData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The `output_structure` output node of the successful calculation if present."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:301,message:"The retrieved temporary folder could not be accessed."},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:303,message:"The retrieved folder did not contain the required XML file."},{status:304,message:"The retrieved folder contained multiple XML files."},{status:305,message:"Both the stdout and XML output files could not be read or parsed."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:320,message:"The XML output file could not be read."},{status:321,message:"The XML output file could not be parsed."},{status:322,message:"The XML output file has an unsupported format."},{status:340,message:"The calculation stopped prematurely because it ran out of walltime but the job was killed by the scheduler before the files were safely written to disk for a potential restart."},{status:350,message:"The parser raised an unexpected exception: {exception}"},{status:360,message:"The code failed in finding a valid reciprocal lattice vector."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:410,message:"The electronic minimization cycle did not reach self-consistency."},{status:461,message:"The code failed with negative dexx in the exchange calculation."},{status:462,message:"The code failed during the cholesky factorization."},{status:463,message:"Too many bands failed to converge during the diagonalization."},{status:464,message:"The S matrix was found to be not positive definite."},{status:465,message:"The `zhegvd` failed in the PPCG diagonalization."},{status:466,message:"The `[Q, R] = qr(X, 0)` failed in the PPCG diagonalization."},{status:467,message:"The eigenvector failed to converge."},{status:468,message:"The factorization in the Broyden routine failed."},{status:481,message:'The k-point parallelization "npools" is too high, some nodes have no k-points.'},{status:500,message:"The ionic minimization cycle did not converge for the given thresholds."},{status:501,message:"Then ionic minimization cycle converged but the thresholds are exceeded in the final SCF."},{status:502,message:"The ionic minimization cycle did not converge after the maximum number of steps."},{status:503,message:"The ionic minimization cycle did not finish because the calculation was interrupted but a partial trajectory and output structure was successfully parsed which can be used for a restart."},{status:510,message:"The electronic minimization cycle failed during an ionic minimization cycle."},{status:511,message:"The ionic minimization cycle converged, but electronic convergence was not reached in the final SCF."},{status:520,message:"The ionic minimization cycle terminated prematurely because of two consecutive failures in the BFGS algorithm."},{status:521,message:"The ionic minimization cycle terminated prematurely because of two consecutive failures in the BFGS algorithm and electronic convergence failed in the final SCF."},{status:531,message:"The electronic minimization cycle did not reach self-consistency."},{status:541,message:"The variable cell optimization broke the symmetry of the k-points."},{status:542,message:"The cell relaxation caused a significant volume contraction and there is not enough space allocated for radial FFT."},{status:710,message:"The electronic minimization cycle did not reach self-consistency, but `scf_must_converge` is `False` and/or `electron_maxstep` is 0."}]},class:"aiida_quantumespresso.calculations.pwimmigrant:PwimmigrantCalculation"},"quantumespresso.q2r":{description:["`CalcJob` implementation for the q2r.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parent_folder",required:!0,valid_types:"RemoteData, FolderData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"force_constants",required:!0,valid_types:"ForceConstantsData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:330,message:"The force constants file could not be read."}]},class:"aiida_quantumespresso.calculations.q2r:Q2rCalculation"},"quantumespresso.seekpath_structure_analysis":{description:["Primitivize the structure with SeeKpath and generate the high symmetry k-point path through its Brillouin zone.",""," This calcfunction will take a structure and pass it through SeeKpath to get the normalized primitive cell and the"," path of high symmetry k-points through its Brillouin zone. Note that the returned primitive cell may differ from the"," original structure in which case the k-points are only congruent with the primitive cell.",""," The keyword arguments can be used to specify various Seekpath parameters, such as:",""," with_time_reversal: True"," reference_distance: 0.025"," recipe: 'hpkot'"," threshold: 1e-07"," symprec: 1e-05"," angle_tolerance: -1.0",""," Note that exact parameters that are available and their defaults will depend on your Seekpath version."],spec:{inputs:[{name:"structure",required:!0,valid_types:"Data",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_quantumespresso.calculations.functions.seekpath_structure_analysis:seekpath_structure_analysis"},"quantumespresso.xspectra":{description:["CalcJob implementation for the xspectra.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"core_wfc_data",required:!0,valid_types:"SinglefileData",info:"Core wavefunction data, generated by the upf2plotcore.sh utility"},{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The K-point sampling to be used for the XSpectra calculation"},{name:"parent_folder",required:!0,valid_types:"RemoteData",info:""},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"gamma_file",required:!1,valid_types:"SinglefileData, NoneType",info:"An optional file containing the data for the broadening function used when `gamma_mode=file`"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:"Parameters for the namelists in the input file."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Use an additional node for special settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"spectra",required:!0,valid_types:"XyData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:302,message:"The retrieved folder did not contain the required stdout output file."},{status:310,message:"An exception was raised while reading the `stdout` file: {exception}"},{status:311,message:"An exception was raised while parsing the `stdout` file: {exception}"},{status:312,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:313,message:"xiabs was set incorrectly, check and ensure that the index value correctly refers to the atomic species containing the core-hole (where the index starts from 1)."},{status:314,message:"xiabs was either set to 0 or less, or was greater than ntyp."},{status:330,message:"The xspectra output file could not be read from the retrieved folder."},{status:331,message:"The spectrum data file could not be read using NumPy genfromtxt"},{status:400,message:"The time limit set for the calculation was exceeded, and the job wrote a save file before exiting."}]},class:"aiida_quantumespresso.calculations.xspectra:XspectraCalculation"}},"aiida.data":{"quantumespresso.force_constants":"aiida_quantumespresso.data.force_constants:ForceConstantsData","quantumespresso.hubbard_structure":"aiida_quantumespresso.data.hubbard_structure:HubbardStructureData"},"aiida.parsers":{"quantumespresso.cp":"aiida_quantumespresso.parsers.cp:CpParser","quantumespresso.dos":"aiida_quantumespresso.parsers.dos:DosParser","quantumespresso.matdyn":"aiida_quantumespresso.parsers.matdyn:MatdynParser","quantumespresso.neb":"aiida_quantumespresso.parsers.neb:NebParser","quantumespresso.open_grid":"aiida_quantumespresso.parsers.open_grid:OpenGridParser","quantumespresso.ph":"aiida_quantumespresso.parsers.ph:PhParser","quantumespresso.pp":"aiida_quantumespresso.parsers.pp:PpParser","quantumespresso.projwfc":"aiida_quantumespresso.parsers.projwfc:ProjwfcParser","quantumespresso.pw":"aiida_quantumespresso.parsers.pw:PwParser","quantumespresso.pw2gw":"aiida_quantumespresso.parsers.pw2gw:Pw2gwParser","quantumespresso.pw2wannier90":"aiida_quantumespresso.parsers.pw2wannier90:Pw2wannier90Parser","quantumespresso.q2r":"aiida_quantumespresso.parsers.q2r:Q2rParser","quantumespresso.xspectra":"aiida_quantumespresso.parsers.xspectra:XspectraParser"},"aiida.tools.calculations":{"quantumespresso.pw":"aiida_quantumespresso.tools.calculations.pw:PwCalculationTools"},"aiida.tools.data.orbitals":{noncollinearhydrogen:"aiida_quantumespresso.tools.data.orbital.noncollinearhydrogen:NoncollinearHydrogenOrbital",spinorbithydrogen:"aiida_quantumespresso.tools.data.orbital.spinorbithydrogen:SpinorbitHydrogenOrbital"},"aiida.workflows":{"quantumespresso.matdyn.base":{description:["Workchain to run a Quantum ESPRESSO matdyn.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"matdyn",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"output_phonon_bands",required:!0,valid_types:"BandsData",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed with an unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_quantumespresso.workflows.matdyn.base:MatdynBaseWorkChain"},"quantumespresso.pdos":{description:["A WorkChain to compute Total & Partial Density of States of a structure, using Quantum Espresso."],spec:{inputs:[{name:"dos",required:!0,valid_types:"Data",info:"Input parameters for the `dos.x` calculation. Note that the `Emin`, `Emax` and `DeltaE` values have to match with those in the `projwfc` inputs."},{name:"nscf",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` of the `nscf` calculation."},{name:"projwfc",required:!0,valid_types:"Data",info:"Input parameters for the `projwfc.x` calculation. Note that the `Emin`, `Emax` and `DeltaE` values have to match with those in the `dos` inputs."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"align_to_fermi",required:!1,valid_types:"Bool",info:"If true, Emin=>Emin-Efermi & Emax=>Emax-Efermi, where Efermi is taken from the `nscf` calculation. Note that it only makes sense to align `Emax` and `Emin` to the fermi level in case they are actually provided by in the `dos` and `projwfc` inputs, since otherwise the "},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If ``True``, work directories of all called calculation will be cleaned at the end of execution."},{name:"dry_run",required:!1,valid_types:"Bool, NoneType",info:"Terminate workchain steps before submitting calculations (test purposes only)."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"scf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` of the `scf` calculation."},{name:"serial_clean",required:!1,valid_types:"Bool, NoneType",info:"If ``True``, calculations will be run in serial, and work directories will be cleaned before the next step."}],outputs:[{name:"dos",required:!0,valid_types:"",info:""},{name:"nscf",required:!0,valid_types:"",info:""},{name:"projwfc",required:!0,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:202,message:"Neither the `kpoints` nor the `kpoints_distance` input was specified for base or nscf namespaces."},{status:401,message:"the SCF sub process failed"},{status:402,message:"the NSCF sub process failed"},{status:403,message:"the DOS sub process failed"},{status:404,message:"the PROJWFC sub process failed"},{status:404,message:"both the DOS and PROJWFC sub process failed"}]},class:"aiida_quantumespresso.workflows.pdos:PdosWorkChain"},"quantumespresso.ph.base":{description:["Workchain to run a Quantum ESPRESSO ph.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"ph",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"only_initialization",required:!1,valid_types:"Bool",info:""},{name:"qpoints",required:!1,valid_types:"KpointsData, NoneType",info:"An explicit qpoints list or mesh. Either this or `qpoints_distance` should to be provided."},{name:"qpoints_distance",required:!1,valid_types:"Float, NoneType",info:"The minimum desired distance in 1/Å between qpoints in reciprocal space. The explicit qpoints will be generated automatically by a calculation function based on the input structure."},{name:"qpoints_force_parity",required:!1,valid_types:"Bool, NoneType",info:"Optional input when constructing the qpoints based on a desired `qpoints_distance`. Setting this to `True` will force the qpoint mesh to have an even number of points along each lattice vector except for any non-periodic directions."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:204,message:"The `metadata.options` did not specify both `resources.num_machines` and `max_wallclock_seconds`. This exit status has been deprecated as the check it corresponded to was incorrect."},{status:300,message:"The calculation failed with an unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:401,message:"The work chain failed to merge the q-points data from multiple `PhCalculation`s because not all q-points were parsed."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_quantumespresso.workflows.ph.base:PhBaseWorkChain"},"quantumespresso.pw.bands":{description:["Workchain to compute a band structure for a given structure using Quantum ESPRESSO pw.x.",""," The logic for the computation of various parameters for the BANDS step is as follows:",""," Number of bands:"," One can specify the number of bands to be used in the BANDS step either directly through the input parameters"," `bands.pw.parameters.SYSTEM.nbnd` or through `nbands_factor`. Note that specifying both is not allowed. When"," neither is specified nothing will be set by the work chain and the default of Quantum ESPRESSO will end up being"," used. If the `nbands_factor` is specified the maximum value of the following values will be used:",""," * `nbnd` of the preceding SCF calculation"," * 0.5 * nelectrons * nbands_factor"," * 0.5 * nelectrons + 4",""," Kpoints:"," There are three options; specify either an existing `KpointsData` through `bands_kpoints`, or specify the"," `bands_kpoint_distance`, or specify neither. For the former those exact kpoints will be used for the BANDS step."," In the two other cases, the structure will first be normalized using SeekPath and the path along high-symmetry"," k-points will be generated on that structure. The distance between kpoints for the path will be equal to that"," of `bands_kpoints_distance` or the SeekPath default if not specified."],spec:{inputs:[{name:"bands",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the BANDS calculation."},{name:"scf",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the SCF calculation."},{name:"structure",required:!0,valid_types:"StructureData",info:"The inputs structure."},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Explicit kpoints to use for the BANDS calculation. Specify either this or `bands_kpoints_distance`."},{name:"bands_kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"Minimum kpoints distance for the BANDS calculation. Specify either this or `bands_kpoints`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nbands_factor",required:!1,valid_types:"Float, NoneType",info:"The number of bands for the BANDS calculation is that used for the SCF multiplied by this factor."},{name:"relax",required:!1,valid_types:"Data",info:"Inputs for the `PwRelaxWorkChain`, if not specified at all, the relaxation step is skipped."}],outputs:[{name:"band_parameters",required:!0,valid_types:"Dict",info:"The output parameters of the BANDS `PwBaseWorkChain`."},{name:"band_structure",required:!0,valid_types:"BandsData",info:"The computed band structure."},{name:"scf_parameters",required:!0,valid_types:"Dict",info:"The output parameters of the SCF `PwBaseWorkChain`."},{name:"primitive_structure",required:!1,valid_types:"StructureData",info:"The normalized and primitivized structure for which the bands are computed."},{name:"seekpath_parameters",required:!1,valid_types:"Dict",info:"The parameters used in the SeeKpath call to normalize the input or relaxed structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"Cannot specify both `nbands_factor` and `bands.pw.parameters.SYSTEM.nbnd`."},{status:202,message:"Cannot specify both `bands_kpoints` and `bands_kpoints_distance`."},{status:401,message:"The PwRelaxWorkChain sub process failed"},{status:402,message:"The scf PwBasexWorkChain sub process failed"},{status:403,message:"The bands PwBasexWorkChain sub process failed"}]},class:"aiida_quantumespresso.workflows.pw.bands:PwBandsWorkChain"},"quantumespresso.pw.base":{description:["Workchain to run a Quantum ESPRESSO pw.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"pw",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"An explicit k-points list or mesh. Either this or `kpoints_distance` has to be provided."},{name:"kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"The minimum desired distance in 1/Å between k-points in reciprocal space. The explicit k-points will be generated automatically by a calculation function based on the input structure."},{name:"kpoints_force_parity",required:!1,valid_types:"Bool, NoneType",info:"Optional input when constructing the k-points based on a desired `kpoints_distance`. Setting this to `True` will force the k-point mesh to have an even number of points along each lattice vector except for any non-periodic directions."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The `output_parameters` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_atomic_occupations",required:!1,valid_types:"Dict",info:""},{name:"output_band",required:!1,valid_types:"BandsData",info:"The `output_band` output node of the successful calculation if present."},{name:"output_kpoints",required:!1,valid_types:"KpointsData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The `output_structure` output node of the successful calculation if present."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"The explicit `pseudos` or `pseudo_family` could not be used to get the necessary pseudos."},{status:202,message:"Neither the `kpoints` nor the `kpoints_distance` input was specified."},{status:203,message:"Neither the `options` nor `automatic_parallelization` input was specified. This exit status has been deprecated as the check it corresponded to was incorrect."},{status:204,message:"The `metadata.options` did not specify both `resources.num_machines` and `max_wallclock_seconds`. This exit status has been deprecated as the check it corresponded to was incorrect."},{status:210,message:"Required key for `automatic_parallelization` was not specified.This exit status has been deprecated as the automatic parallellization feature was removed."},{status:211,message:"Unrecognized keys were specified for `automatic_parallelization`.This exit status has been deprecated as the automatic parallellization feature was removed."},{status:300,message:"The calculation failed with an unidentified unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:310,message:"The calculation failed with a known unrecoverable error."},{status:320,message:"The initialization calculation failed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."},{status:501,message:"Then ionic minimization cycle converged but the thresholds are exceeded in the final SCF."},{status:710,message:"The electronic minimization cycle did not reach self-consistency, but `scf_must_converge` is `False` and/or `electron_maxstep` is 0."}]},class:"aiida_quantumespresso.workflows.pw.base:PwBaseWorkChain"},"quantumespresso.pw.relax":{description:["Workchain to relax a structure using Quantum ESPRESSO pw.x."],spec:{inputs:[{name:"base",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the main relax loop."},{name:"structure",required:!0,valid_types:"StructureData",info:"The inputs structure."},{name:"base_final_scf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the final scf."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"max_meta_convergence_iterations",required:!1,valid_types:"Int",info:"The maximum number of variable cell relax iterations in the meta convergence cycle."},{name:"meta_convergence",required:!1,valid_types:"Bool",info:"If `True` the workchain will perform a meta-convergence on the cell volume."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"volume_convergence",required:!1,valid_types:"Float",info:"The volume difference threshold between two consecutive meta convergence iterations."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The `output_parameters` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"output_atomic_occupations",required:!1,valid_types:"Dict",info:""},{name:"output_band",required:!1,valid_types:"BandsData",info:"The `output_band` output node of the successful calculation if present."},{name:"output_kpoints",required:!1,valid_types:"KpointsData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:"The successfully relaxed structure."},{name:"output_trajectory",required:!1,valid_types:"TrajectoryData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"the relax PwBaseWorkChain sub process failed"},{status:402,message:"the final scf PwBaseWorkChain sub process failed"}]},class:"aiida_quantumespresso.workflows.pw.relax:PwRelaxWorkChain"},"quantumespresso.q2r.base":{description:["Workchain to run a Quantum ESPRESSO q2r.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"q2r",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"force_constants",required:!0,valid_types:"ForceConstantsData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed with an unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_quantumespresso.workflows.q2r.base:Q2rBaseWorkChain"},"quantumespresso.xps":{description:["Workchain to compute X-ray photoelectron spectra (XPS) for a given structure.",""," The WorkChain itself firstly calls the PwRelaxWorkChain to relax the input structure if"," required. Then determines the input settings for each XPS calculation automatically using"," ``get_xspectra_structures()``. The input structures are generated from the standardized"," structure by converting each to a supercell with cell dimensions of at least 8.0 angstrom"," in each periodic dimension in order to sufficiently reduce the unphysical interaction"," of the core-hole with neighbouring images. The size of the minimum size requirement can be"," overriden by the user if required. Then the standard Delta-Self-Consistent-Field (ΔSCF)"," method is used to get the XPS binding energy. Finally, the XPS spectrum is calculated"," using the Voigt profile."],spec:{inputs:[{name:"ch_scf",required:!0,valid_types:"Data",info:"Input parameters for the basic xps workflow (core-hole SCF)."},{name:"core_hole_pseudos",required:!0,valid_types:"UpfData, UpfData",info:'Dynamic namespace for pairs of excited-state pseudopotentials for each absorbing element. Must use the mapping "{element}" : {Upf}".'},{name:"gipaw_pseudos",required:!0,valid_types:"UpfData, UpfData",info:'Dynamic namespace for pairs of ground-state pseudopotentials for each absorbing element. Must use the mapping "{element}" : {Upf}".'},{name:"structure",required:!0,valid_types:"StructureData",info:"Structure to be used for calculation."},{name:"abs_atom_marker",required:!1,valid_types:"Str",info:"The name for the Kind representing the absorbing atom in the structure. Will be used in all structures generated in ``get_xspectra_structures`` step."},{name:"atoms_list",required:!1,valid_types:"List, NoneType",info:"The indices of atoms to be considered for analysis."},{name:"calc_binding_energy",required:!1,valid_types:"Bool",info:"If `True`, run scf calculation for the supercell."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculations will be cleaned at the end of execution."},{name:"core_hole_treatments",required:!1,valid_types:"Dict, NoneType",info:"Optional dictionary to set core-hole treatment to all elements present. The default full-core-hole treatment will be used if not specified."},{name:"correction_energies",required:!1,valid_types:"Dict, NoneType",info:"Optional dictionary to set the correction energy to all elements present. "},{name:"dry_run",required:!1,valid_types:"Bool, NoneType",info:"Terminate workchain steps before submitting calculations (test purposes only)."},{name:"elements_list",required:!1,valid_types:"List, NoneType",info:"The list of elements to be considered for analysis, each must be valid elements of the periodic table."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax",required:!1,valid_types:"Data",info:"Input parameters for the relax process. If not specified at all, the relaxation step is skipped."},{name:"spglib_settings",required:!1,valid_types:"Dict, NoneType",info:"Optional settings dictionary for the spglib call within ``get_xspectra_structures``."},{name:"structure_preparation_settings",required:!1,valid_types:"Dict, Float, Int, Bool, Str",info:"Optional settings dictionary for the ``get_xspectra_structures()`` method."},{name:"voight_gamma",required:!1,valid_types:"Float",info:"The gamma parameter for the Lorenzian broadening in the Voight method."},{name:"voight_sigma",required:!1,valid_types:"Float",info:"The sigma parameter for the gaussian broadening in the Voight method."}],outputs:[{name:"binding_energies",required:!0,valid_types:"Dict",info:"All the binding energy values for each element calculated by the WorkChain."},{name:"chemical_shifts",required:!0,valid_types:"Dict",info:"All the chemical shift values for each element calculated by the WorkChain."},{name:"final_spectra_be",required:!0,valid_types:"XyData",info:"The fully-resolved spectra for each element based on binding energy."},{name:"final_spectra_cls",required:!0,valid_types:"XyData",info:"The fully-resolved spectra for each element based on chemical shift."},{name:"output_parameters_ch_scf",required:!0,valid_types:"Dict",info:"The output parameters of each ``PwBaseWorkChain`` performed``."},{name:"optimized_structure",required:!1,valid_types:"StructureData",info:"The optimized structure from the ``relax`` process."},{name:"output_parameters_relax",required:!1,valid_types:"Dict",info:"The output_parameters of the relax step."},{name:"output_parameters_scf",required:!1,valid_types:"Dict",info:"The output_parameters of the scf step."},{name:"standardized_structure",required:!1,valid_types:"StructureData",info:"The standardized crystal structure used to generate structures for XPS sub-processes."},{name:"supercell_structure",required:!1,valid_types:"StructureData",info:"The supercell of ``outputs.standardized_structure`` used to generate structures for XPS sub-processes."},{name:"symmetry_analysis_data",required:!1,valid_types:"Dict",info:"The output parameters from ``get_xspectra_structures()``."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"The Relax sub process failed"},{status:402,message:"The SCF Pw sub processes failed"},{status:402,message:"One or more CH_SCF Pw sub processes failed"}]},class:"aiida_quantumespresso.workflows.xps:XpsWorkChain"},"quantumespresso.xspectra.base":{description:["Workchain to run a Quantum ESPRESSO xspectra.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"xspectra",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"An explicit k-points mesh. Either this or `kpoints_distance` has to be provided."},{name:"kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"The minimum desired distance in 1/Å between k-points in reciprocal space. The explicit k-points will be generated automatically by a calculation function based on the input structure."},{name:"kpoints_force_parity",required:!1,valid_types:"Bool, NoneType",info:"Optional input when constructing the k-points based on a desired `kpoints_distance`. Setting this to `True` will force the k-point mesh to have an even number of points along each lattice vector except for any non-periodic directions."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"spectra",required:!0,valid_types:"XyData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:202,message:"Neither the `kpoints` nor the `kpoints_distance` input was specified."},{status:300,message:"The calculation failed with an unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_quantumespresso.workflows.xspectra.base:XspectraBaseWorkChain"},"quantumespresso.xspectra.core":{description:["Workchain to compute X-ray absorption spectra for a given structure using Quantum ESPRESSO.",""," The workflow follows the process required to compute the XAS of an input structure: an SCF calculation is performed"," using the provided structure, which is then followed by the calculation of the XAS itself by XSpectra. The"," calculations performed by the WorkChain in a typical run will be:",""," - PwSCF calculation with pw.x of the input structure with a core-hole present."," - Generation of core-wavefunction data with upf2plotcore.sh (if requested)."," - XAS calculation with xspectra.x to compute the Lanczos coefficients and print the XANES spectra for the"," polarisation vectors requested in the input."," - Collation of output data from pw.x and xspectra.x calculations, including a combination of XANES dipole spectra"," based on polarisation vectors to represent the powder spectrum of the structure (if requested).",""," If ``run_replot = True`` is set in the inputs (defaults to False), the WorkChain will run a second xspectra.x"," calculation which replots the spectra produced from the ``xs_prod`` step. This option can be very useful for"," obtaining a final spectrum at low levels of broadening (relative to the default of 0.5 eV), particularly as higher"," levels of broadening significantly speed up the convergence of the Lanczos procedure. Inputs for the replot"," calculation are found in the ``xs_plot`` namespace.",""," The core-wavefunction plot derived from the ground-state of the absorbing element can be provided as a top-level"," input or produced by the WorkChain. If left to the WorkChain, the ground-state pseudopotential assigned to the"," absorbing element will be used to generate this data using the upf2plotcore.sh utility script (via the"," ``aiida-shell`` plugin).",""," In its current stage of development, the workflow requires the following:",""," - An input structure where the desired absorbing atom in the system is marked as a separate Kind. The default"," behaviour for the WorkChain is to set the Kind name as 'X', however this can be changed via the `overrides`"," dictionary."," - A code node for ``upf2plotcore``, configured for the ``aiida-shell`` plugin"," (https://github.com/sphuber/aiida-shell). Alternatively, a ``SinglefileData`` node from a previous ``ShellJob``"," run can be supplied under ``inputs.core_wfc_data``."," - A suitable pair of pseudopotentials for the element type of the absorbing atom, one for the ground-state occupancy"," which contains GIPAW informtation for the core level of interest for the XAS (e.g. 1s in the case of a K-edge"," calculation) and the other containing a core hole. (For the moment this can be passed either via the"," ``core_hole_pseudos`` field in ``get_builder_from_protocol`` or via the overrides, but will be changed later once"," full families of core-hole pseudopotentials become available)."],spec:{inputs:[{name:"eps_vectors",required:!0,valid_types:"List",info:"The list of 3-vectors to use in XSpectra sub-processes. The number of sub-lists will subsequently define the number of XSpectra calculations to perform"},{name:"scf",required:!0,valid_types:"Data",info:"Input parameters for the `pw.x` calculation."},{name:"structure",required:!0,valid_types:"StructureData, HubbardStructureData",info:"Structure to be used for calculation, with at least one site containing the `abs_atom_marker` as the kind label."},{name:"xs_prod",required:!0,valid_types:"Data",info:"Input parameters for the `xspectra.x` calculation to compute the Lanczos."},{name:"abs_atom_marker",required:!1,valid_types:"Str, NoneType",info:"The name for the Kind representing the absorbing atom in the structure. Must corespond to a Kind within the StructureData node supplied to the calculation."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"core_wfc_data",required:!1,valid_types:"SinglefileData, NoneType",info:"The core wavefunction data file extracted from the ground-state pseudo for the absorbing atom."},{name:"dry_run",required:!1,valid_types:"Bool, NoneType",info:"Terminate workchain steps before submitting calculations (test purposes only)."},{name:"get_powder_spectrum",required:!1,valid_types:"Bool",info:"If `True`, the WorkChain will combine XANES dipole spectra computed using the XAS basis vectors defined according to the `get_powder_spectrum` CalcFunction."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"run_replot",required:!1,valid_types:"Bool",info:""},{name:"upf2plotcore_code",required:!1,valid_types:"AbstractCode, NoneType",info:"The code node required for upf2plotcore.sh configured for ``aiida-shell``. Must be provided if `core_wfc_data` is not provided."},{name:"xs_plot",required:!1,valid_types:"Data",info:"Input parameters for the re-plot `xspectra.x` calculation of the Lanczos."}],outputs:[{name:"parameters_scf",required:!0,valid_types:"Dict",info:"The output parameters of the SCF `PwBaseWorkChain`."},{name:"parameters_xspectra",required:!0,valid_types:"Dict",info:"The output dictionaries of each `XspectraBaseWorkChain` performed"},{name:"spectra",required:!0,valid_types:"XyData",info:"An XyData node containing all the final spectra produced by the WorkChain."},{name:"powder_spectrum",required:!1,valid_types:"XyData",info:"The simulated powder spectrum"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"The SCF sub process failed"},{status:402,message:"One or more XSpectra sub processes failed"},{status:403,message:"The pseudo for the absorbing element contains no GIPAW information."}]},class:"aiida_quantumespresso.workflows.xspectra.core:XspectraCoreWorkChain"},"quantumespresso.xspectra.crystal":{description:["Workchain to compute all X-ray absorption spectra for a given structure using Quantum ESPRESSO.",""," The WorkChain follows the process required to compute all the K-edge XAS spectra for each"," element in a given structure. The WorkChain itself firstly calls the PwRelaxWorkChain to"," relax the input structure, then determines the input settings for each XAS"," calculation automatically using ``get_xspectra_structures()``:",""," - Firstly the input structure is converted to its conventional standard cell using"," ``spglib`` and detects the space group number for the conventional cell."," - Symmetry analysis of the standardized structure using ``spglib`` is then used to"," determine the number of non-equivalent atomic sites in the structure for each"," element considered for analysis.",""," Using the symmetry data returned from ``get_xspectra_structures``, input structures for"," the XspectraCoreWorkChain are generated from the standardized structure by converting each"," to a supercell with cell dimensions of at least 8.0 angstroms in each periodic dimension -"," required in order to sufficiently reduce the unphysical interaction of the core-hole with"," neighbouring images. The size of the minimum size requirement can be overriden by the"," user if required. The WorkChain then uses the space group number to set the list of"," polarisation vectors for the ``XspectraCoreWorkChain`` to compute for all subsequent"," calculations."],spec:{inputs:[{name:"core",required:!0,valid_types:"Data",info:"Input parameters for the basic xspectra workflow (core-hole SCF + XAS."},{name:"core_hole_pseudos",required:!0,valid_types:"UpfData, UpfData",info:'Dynamic namespace for pairs of excited-state pseudopotentials for each absorbing element. Must use the mapping "{element}" : {Upf}".'},{name:"elements_list",required:!0,valid_types:"List",info:"The list of elements to be considered for analysis, each must be a valid element of the periodic table."},{name:"gipaw_pseudos",required:!0,valid_types:"UpfData, UpfData",info:'Dynamic namespace for pairs of ground-state pseudopotentials for each absorbing element. Must use the mapping "{element}" : {Upf}.'},{name:"structure",required:!0,valid_types:"StructureData",info:"Structure to be used for calculation."},{name:"abs_atom_marker",required:!1,valid_types:"Str",info:"The name for the Kind representing the absorbing atom in the structure. Will be used in all structures generated in ``get_xspectra_structures`` step."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculations will be cleaned at the end of execution."},{name:"core_hole_treatments",required:!1,valid_types:"Dict, NoneType",info:"Optional dictionary to set core-hole treatment to given elements present. The default full-core-hole treatment will be used if not specified."},{name:"core_wfc_data",required:!1,valid_types:"SinglefileData",info:"Input namespace to provide core wavefunction inputs for each element. Must follow the format: ``core_wfc_data__{symbol} = {node}``"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax",required:!1,valid_types:"Data",info:"Input parameters for the relax process. If not specified at all, the relaxation step is skipped."},{name:"return_all_powder_spectra",required:!1,valid_types:"Bool",info:"If ``True``, the WorkChain will return all ``powder_spectrum`` nodes from each ``XspectraCoreWorkChain`` sub-process."},{name:"spglib_settings",required:!1,valid_types:"Dict, NoneType",info:"Optional settings dictionary for the spglib call within ``get_xspectra_structures``."},{name:"structure_preparation_settings",required:!1,valid_types:"Dict, Float, Int, Bool, Str",info:"Optional settings dictionary for the ``get_xspectra_structures()`` method."},{name:"upf2plotcore_code",required:!1,valid_types:"AbstractCode, NoneType",info:"Code node for the upf2plotcore.sh ShellJob code."}],outputs:[{name:"final_spectra",required:!0,valid_types:"XyData",info:"The fully-resolved spectra for each element"},{name:"supercell_structure",required:!0,valid_types:"StructureData",info:"The supercell of ``outputs.standardized_structure`` used to generate structures for XSpectra sub-processes."},{name:"symmetry_analysis_data",required:!0,valid_types:"Dict",info:"The output parameters from ``get_xspectra_structures()``."},{name:"optimized_structure",required:!1,valid_types:"StructureData",info:"The optimized structure from the ``relax`` process."},{name:"parameters_relax",required:!1,valid_types:"Dict",info:"The output_parameters of the relax step."},{name:"parameters_scf",required:!1,valid_types:"Dict",info:"The output parameters of each ``PwBaseWorkChain`` performed in each ``XspectraCoreWorkChain``."},{name:"parameters_xspectra",required:!1,valid_types:"Dict",info:"The output dictionaries of each `XspectraCalculation` performed"},{name:"powder_spectra",required:!1,valid_types:"XyData",info:"All the spectra generated by the WorkChain."},{name:"standardized_structure",required:!1,valid_types:"StructureData",info:"The standardized crystal structure used to generate structures for XSpectra sub-processes."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:401,message:"The Relax sub process failed"},{status:402,message:"One or more XSpectra workflows failed"},{status:403,message:"The pseudos for one or more absorbing elements contain no GIPAW information."}]},class:"aiida_quantumespresso.workflows.xspectra.crystal:XspectraCrystalWorkChain"}},console_scripts:{"aiida-quantumespresso":"aiida_quantumespresso.cli:cmd_root"}},commits_count:33,development_status:"stable",errors:[],warnings:["W010: Entry point 'noncollinearhydrogen' does not start with prefix 'quantumespresso.'","W010: Entry point 'spinorbithydrogen' does not start with prefix 'quantumespresso.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:20},{colorclass:"brown",text:"Parsers",count:13},{colorclass:"red",text:"Data",count:2},{colorclass:"green",text:"Workflows",count:11},{colorclass:"purple",text:"Console scripts",count:1},{colorclass:"orange",text:"Other (Tools calculations, Tools data orbitals)",count:3}],pip_install_cmd:"pip install aiida-quantumespresso",is_installable:"True"},"aiida-quantumespresso-hp":{code_home:"https://github.com/aiidateam/aiida-quantumespresso-hp",entry_point_prefix:"quantumespresso.hp",pip_url:"git+https://github.com/aiidateam/aiida-quantumespresso-hp",name:"aiida-quantumespresso-hp",package_name:"aiida_quantumespresso_hp",hosted_on:"github.com",metadata:{classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: POSIX :: Linux","Operating System :: MacOS :: MacOS X","Programming Language :: Python","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11"],author:"Sebastiaan P. Huber",author_email:"mail@sphuber.net"},aiida_version:"~=2.2",entry_points:{"aiida.calculations":{"quantumespresso.hp":{description:["`CalcJob` implementation for the hp.x code of Quantum ESPRESSO."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"The input parameters for the namelists."},{name:"parent_hp",required:!0,valid_types:"FolderData",info:""},{name:"parent_scf",required:!0,valid_types:"RemoteData",info:""},{name:"qpoints",required:!0,valid_types:"KpointsData",info:"The q-point grid on which to perform the perturbative calculation."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Optional node for special settings."}],outputs:[{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"hubbard",required:!1,valid_types:"Dict",info:"Parsed Hubbard parameters from the ``Hubbard_parameters.dat`` file."},{name:"hubbard_chi",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_matrices",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData",info:"``HubbardStructureData`` containing the new Hubbard parameters."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"The retrieved folder data node could not be accessed."},{status:210,message:"The retrieved folder did not contain the required stdout output file."},{status:211,message:"The retrieved folder did not contain the required hubbard output file."},{status:212,message:"The retrieved folder did not contain the required hubbard chi output file."},{status:300,message:"Problems with one or more output files."},{status:310,message:"The stdout output file could not be read."},{status:311,message:"The stdout output file could not be parsed."},{status:312,message:"The stdout output file was incomplete."},{status:313,message:"The `HUBBARD.dat` could not be parsed."},{status:350,message:"The namelist in the input file contained invalid syntax and could not be parsed."},{status:360,message:"One of the required perturbation inputs files was not found."},{status:365,message:"The atomic positions were not sorted with Hubbard sites first."},{status:366,message:"The code failed due to Fermi shift, probably due to low energy cutoff or due to an incorrect treatment of an insulating state (i.e. no smearing shoudl be used)."},{status:400,message:"The calculation stopped prematurely because it ran out of walltime."},{status:410,message:"The electronic minimization cycle did not reach self-consistency."},{status:462,message:"The code failed during the cholesky factorization."},{status:490,message:"The code failed to reconstruct the full chi matrix as some chi matrices were missing"},{status:495,message:"The code failed due to incompatibility between the FFT grid and the parallelization options."}]},class:"aiida_quantumespresso_hp.calculations.hp:HpCalculation"}},"aiida.parsers":{"quantumespresso.hp":"aiida_quantumespresso_hp.parsers.hp:HpParser"},"aiida.workflows":{"quantumespresso.hp.main":{description:["Work chain to run a Quantum ESPRESSO hp.x calculation.",""," If the `parallelize_atoms` input is set to `True`, the calculation will be parallelized over the Hubbard atoms by"," running the `HpParallelizeAtomsWorkChain`. When parallelizing over atoms, if the `parallelize_qpoints` is `True`,"," each `HpParallelizeAtomsWorkChain` will be parallelized over its perturbations (q points) running the"," `HpParallelizeQpointsWorkChain`. Otherwise a single `HpBaseWorkChain` will be launched that will compute"," every Hubbard atom, and every q point in serial.",""," .. important:: q point parallelization is only possible when parallelization over atoms is performed."],spec:{inputs:[{name:"hp",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_concurrent_base_workchains",required:!1,valid_types:"Int, NoneType",info:""},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parallelize_atoms",required:!1,valid_types:"Bool",info:""},{name:"parallelize_qpoints",required:!1,valid_types:"Bool",info:""},{name:"qpoints",required:!1,valid_types:"KpointsData, NoneType",info:"An explicit q-points list or mesh. Either this or `qpoints_distance` has to be provided."},{name:"qpoints_distance",required:!1,valid_types:"Float, NoneType",info:"The minimum desired distance in 1/Å between q-points in reciprocal space. The explicit q-points will be generated automatically by a calculation function based on the input structure."},{name:"qpoints_force_parity",required:!1,valid_types:"Bool, NoneType",info:"Optional input when constructing the q-points based on a desired `qpoints_distance`. Setting this to `True` will force the q-point mesh to have an even number of points along each lattice vector except for any non-periodic directions."}],outputs:[{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"hubbard",required:!1,valid_types:"Dict",info:"Parsed Hubbard parameters from the ``Hubbard_parameters.dat`` file."},{name:"hubbard_chi",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_matrices",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData",info:"``HubbardStructureData`` containing the new Hubbard parameters."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"Neither the `qpoints` nor the `qpoints_distance`, or the `hp.hubbard_structure` input were specified."},{status:300,message:"A child work chain failed."}]},class:"aiida_quantumespresso_hp.workflows.hp.main:HpWorkChain"},"quantumespresso.hp.parallelize_atoms":{description:["Work chain to launch a Quantum Espresso hp.x calculation parallelizing over the Hubbard atoms."],spec:{inputs:[{name:"hp",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"init_walltime",required:!1,valid_types:"int",info:"The walltime of the initialization `HpBaseWorkChain` in seconds (default: 3600)."},{name:"max_concurrent_base_workchains",required:!1,valid_types:"Int, NoneType",info:""},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parallelize_qpoints",required:!1,valid_types:"Bool",info:""}],outputs:[{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"hubbard",required:!1,valid_types:"Dict",info:"Parsed Hubbard parameters from the ``Hubbard_parameters.dat`` file."},{name:"hubbard_chi",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_matrices",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData",info:"``HubbardStructureData`` containing the new Hubbard parameters."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"A child work chain failed."},{status:301,message:"The child work chain failed."},{status:302,message:"The child work chain failed."}]},class:"aiida_quantumespresso_hp.workflows.hp.parallelize_atoms:HpParallelizeAtomsWorkChain"},"quantumespresso.hp.parallelize_qpoints":{description:["Work chain to launch a Quantum Espresso hp.x calculation parallelizing over the q points on a single Hubbard atom."],spec:{inputs:[{name:"hp",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"init_walltime",required:!1,valid_types:"int",info:"The walltime of the initialization `HpBaseWorkChain` in seconds (default: 3600)."},{name:"max_concurrent_base_workchains",required:!1,valid_types:"Int, NoneType",info:""},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"hubbard",required:!1,valid_types:"Dict",info:"Parsed Hubbard parameters from the ``Hubbard_parameters.dat`` file."},{name:"hubbard_chi",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_matrices",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData",info:"``HubbardStructureData`` containing the new Hubbard parameters."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"A child work chain failed."},{status:301,message:"The child work chain failed."},{status:302,message:"The child work chain failed."}]},class:"aiida_quantumespresso_hp.workflows.hp.parallelize_qpoints:HpParallelizeQpointsWorkChain"},"quantumespresso.hp.base":{description:["Workchain to run a Quantum ESPRESSO hp.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"hp",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"only_initialization",required:!1,valid_types:"Bool",info:""}],outputs:[{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"hubbard",required:!1,valid_types:"Dict",info:"Parsed Hubbard parameters from the ``Hubbard_parameters.dat`` file."},{name:"hubbard_chi",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_matrices",required:!1,valid_types:"ArrayData",info:""},{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData",info:"``HubbardStructureData`` containing the new Hubbard parameters."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"The calculation failed with an unrecoverable error."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_quantumespresso_hp.workflows.hp.base:HpBaseWorkChain"},"quantumespresso.hp.hubbard":{description:["Workchain computing the self-consistent Hubbard parameters of a structure.",""," It iteratively relaxes the structure (optional) with the ``PwRelaxWorkChain``"," and computes the Hubbard parameters through the ``HpWorkChain``,"," using the remote folder of an scf performed via the ``PwBaseWorkChain``,"," until the Hubbard values are converged within certain tolerance(s).",""," The procedure in each step of the convergence cycle is slightly different depending on the electronic and"," magnetic properties of the system. Each cycle will roughly consist of three steps:",""," * Relaxing the structure at the current Hubbard values (optional)."," * One or two SCF calculations depending whether the system is metallic or insulating, respectively."," * A self-consistent calculation of the Hubbard parameters, restarted from the last SCF run.",""," The possible options for the set of SCF calculations that have to be run in the second step look are:",""," * Metals:",""," - SCF with smearing.",""," * Insulators",""," - SCF with smearing."," - SCF with fixed occupations; if magnetic, total magnetization and number of bands"," are fixed to the values found from the previous SCF calculation.",""," When convergence is achieved a node will be returned containing the final converged"," :class:`~aiida_quantumespresso.data.hubbard_structure.HubbardStructureData`."],spec:{inputs:[{name:"hubbard",required:!0,valid_types:"Data",info:""},{name:"hubbard_structure",required:!0,valid_types:"HubbardStructureData",info:"The HubbardStructureData containing the initialized parameters for triggering the Hubbard atoms which the `hp.x` code will perturbe."},{name:"scf",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations of the (relax-)scf-hp cycle."},{name:"meta_convergence",required:!1,valid_types:"Bool",info:"Whether performing the self-consistent cycle. If False, it will stop at the first iteration."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax",required:!1,valid_types:"Data",info:"Inputs for the `PwRelaxWorkChain` that, when defined, will iteratively relax the structure."},{name:"relax_frequency",required:!1,valid_types:"Int, NoneType",info:"Integer value referring to the number of iterations to wait before performing the `relax` step."},{name:"skip_relax_iterations",required:!1,valid_types:"Int, NoneType",info:"The number of iterations for skipping the `relax` step without performing check on parameters convergence."},{name:"tolerance_intersite",required:!1,valid_types:"Float",info:"Tolerance value for self-consistent DFT+U+V calculation. It refers to the only off-diagonal elements V."},{name:"tolerance_onsite",required:!1,valid_types:"Float",info:"Tolerance value for self-consistent calculation of Hubbard U. In case of DFT+U+V calculation, it refers to the diagonal elements (i.e. on-site)."}],outputs:[{name:"hubbard_structure",required:!1,valid_types:"HubbardStructureData",info:"The Hubbard structure containing the structure and associated Hubbard parameters."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:330,message:"Failed to determine the correct pseudo potential after the structure changed its kind names."},{status:401,message:"The reconnaissance PwBaseWorkChain sub process failed"},{status:402,message:"The PwRelaxWorkChain sub process failed in iteration {iteration}"},{status:403,message:"The scf PwBaseWorkChain sub process failed in iteration {iteration}"},{status:404,message:"The HpWorkChain sub process failed in iteration {iteration}"},{status:405,message:"The scf PwBaseWorkChain sub process in iteration {iteration}returned a non integer total magnetization (threshold exceeded)."},{status:601,message:"The Hubbard parameters did not converge at the last iteration #{iteration}"}]},class:"aiida_quantumespresso_hp.workflows.hubbard:SelfConsistentHubbardWorkChain"}}},commits_count:14,development_status:"beta",errors:[],warnings:["W009: Prefix 'quantumespresso.hp' does not follow naming convention."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:5}],pip_install_cmd:"pip install git+https://github.com/aiidateam/aiida-quantumespresso-hp",is_installable:"True"},"aiida-raspa":{code_home:"https://github.com/yakutovicha/aiida-raspa",entry_point_prefix:"raspa",pip_url:"aiida-raspa",plugin_info:"https://raw.github.com/yakutovicha/aiida-raspa/master/setup.json",name:"aiida-raspa",package_name:"aiida_raspa",hosted_on:"github.com",metadata:{release_date:"2023-08-26",description:"AiiDA plugin for RASPA code",author_email:"Aliaksandr Yakutovich , Miriam Pougin ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"2.0.0"},aiida_version:">=2.3,<3.0",entry_points:{"aiida.calculations":{raspa:{description:["This is a RaspaCalculation, subclass of CalcJob, to prepare input for RASPA code."," For information on RASPA, refer to: https://github.com/iraspa/raspa2."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"block_pocket",required:!1,valid_types:"SinglefileData",info:"Zeo++ block pocket file"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"file",required:!1,valid_types:"SinglefileData",info:"Additional input file(s)"},{name:"framework",required:!1,valid_types:"CifData",info:"Input framework(s)"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote folder used to continue the same simulation stating from the binary restarts."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"retrieved_parent_folder",required:!1,valid_types:"FolderData, NoneType",info:"To use an old calculation as a starting poing for a new one."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional input parameters"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The results of a calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"warnings",required:!1,valid_types:"List",info:"Warnings that appeared during the calculation"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The retrieved folder data node could not be accessed."},{status:101,message:"The retrieved folder does not contain an output file."},{status:102,message:'The output does not contain "Starting simulation".'},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:500,message:"The calculation could not be completed due to the lack of time."}]},class:"aiida_raspa.calculations:RaspaCalculation"}},"aiida.parsers":{raspa:"aiida_raspa.parsers:RaspaParser"},"aiida.workflows":{"raspa.base":{description:["Workchain to run a RASPA calculation with automated error handling and restarts."],spec:{inputs:[{name:"raspa",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The results of a calculation"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"warnings",required:!1,valid_types:"List",info:"Warnings that appeared during the calculation"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_raspa.workchains:RaspaBaseWorkChain"}}},commits_count:0,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-raspa",is_installable:"True"},"aiida-shell":{code_home:"https://github.com/sphuber/aiida-shell",entry_point_prefix:"core",pip_url:"aiida-shell",plugin_info:"https://raw.github.com/sphuber/aiida-shell/master/pyproject.toml",name:"aiida-shell",package_name:"aiida_shell",hosted_on:"github.com",metadata:{release_date:"2024-09-18",description:"AiiDA plugin that makes running shell commands easy.",author_email:'"Sebastiaan P. Huber" ',classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering"],version:"0.8.0"},aiida_version:">=2.6.1,<3.0",entry_points:{"aiida.calculations":{"core.shell":{description:["Implementation of :class:`aiida.engine.CalcJob` to run a simple shell command."],spec:{inputs:[{name:"code",required:!0,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"arguments",required:!1,valid_types:"List, NoneType",info:""},{name:"filenames",required:!1,valid_types:"Dict, NoneType",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"nodes",required:!1,valid_types:"Data",info:""},{name:"outputs",required:!1,valid_types:"List, NoneType",info:""},{name:"parser",required:!1,valid_types:"EntryPointData, PickledData, NoneType",info:""},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Exit status could not be determined: exit status file was not retrieved."},{status:301,message:"Exit status could not be determined: exit status file does not contain a valid integer."},{status:302,message:"The stdout file was not retrieved."},{status:303,message:"One or more output files defined in the `outputs` input were not retrieved: {missing_filepaths}."},{status:310,message:"Callable specified in the `parser` input excepted: {exception}."},{status:400,message:"The command exited with a non-zero status: {status} {stderr}."},{status:410,message:"The command exited with a zero status but the stderr was not empty."}]},class:"aiida_shell.calculations.shell:ShellJob"}},"aiida.data":{"core.code.installed.shell":"aiida_shell.data.code:ShellCode","core.entry_point":"aiida_shell.data.entry_point:EntryPointData","core.pickled":"aiida_shell.data.pickled:PickledData"},"aiida.parsers":{"core.shell":"aiida_shell.parsers.shell:ShellParser"}},commits_count:36,development_status:"beta",errors:[],warnings:["W009: Prefix 'core' does not follow naming convention."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:3}],pip_install_cmd:"pip install aiida-shell",is_installable:"True"},"aiida-siesta":{code_home:"https://github.com/siesta-project/aiida_siesta_plugin/tree/master",documentation_url:"https://aiida-siesta-plugin.readthedocs.io/",entry_point_prefix:"siesta",pip_url:"aiida-siesta",name:"aiida-siesta",package_name:"aiida_siesta",hosted_on:"github.com",metadata:{release_date:"2022-07-17",description:"A plugin for Siesta's basic functionality within the AiiDA framework.",author_email:'Albero Garcia , "Victor M. Garcia-Suarez" , Emanuele Bosoni , Vladimir Dikan , Pol Febrer ',classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"2.0.0"},aiida_version:">=2.0.0,<3.0.0",entry_points:{"aiida.calculations":{"siesta.siesta":{description:["Siesta calculator class for AiiDA."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The calculation results"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"Optional band structure"},{name:"forces_and_stress",required:!1,valid_types:"ArrayData",info:"Optional forces and stress"},{name:"ion_files",required:!1,valid_types:"IonData",info:""},{name:"optical_eps2",required:!1,valid_types:"ArrayData",info:"Optional eps2 optical data"},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Optional relaxed structure"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:350,message:'Statement "Job completed" not detected, unknown error'},{status:448,message:"Problems in the polarization of a basis element"},{status:449,message:"Split_norm parameter too small"},{status:450,message:"Calculation did not reach scf convergence!"},{status:451,message:"Calculation did not reach geometry convergence!"},{status:452,message:"Bands analysis was requested, but file is not present"},{status:453,message:"Failure while parsing the bands file"},{status:454,message:"Optical calculation requested, but file is not present"}]},class:"aiida_siesta.calculations.siesta:SiestaCalculation"},"siesta.stm":{description:['Plugin for the "plstm" program in the Siesta distribution.',"",' The "plstm" program takes the .LDOS file and generates a plot file with a simulated STM image.',' It supports both the old "plstm" versions (inputs in a files) and the new ones (inputs in the command',' line). Spin options are supported only in recent "plstm" versions, therefore ignored otherwise.'],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"ldos_folder",required:!0,valid_types:"RemoteData",info:"Parent folder"},{name:"mode",required:!0,valid_types:"Str",info:'Allowed values are "constant-height" or "constant-current"'},{name:"value",required:!0,valid_types:"Float",info:"Value of height in Ang or value of current in e/bohr**3"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"},{name:"spin_option",required:!1,valid_types:"Str",info:'Spin option follows plstm sintax: "q" no spin, "s" total spin, "x","y","z" the three spin components'}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"For the moment only parser version and name of .STM file"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"stm_array",required:!0,valid_types:"ArrayData",info:"The contour data for the STM image"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The retrieved folder data node could not be accessed."},{status:101,message:"The retrieved folder does not contain a CH.STM file"},{status:102,message:"The .STM file can not be read"},{status:102,message:"The array containing the STM data can not be produced"},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_siesta.calculations.stm:STMCalculation"}},"aiida.parsers":{"siesta.parser":"aiida_siesta.parsers.siesta:SiestaParser","siesta.stm":"aiida_siesta.parsers.stm:STMParser"},"aiida.workflows":{"siesta.base":{description:["Base Workchain to launch a total energy calculation via Siesta."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The calculation results"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"Optional band structure"},{name:"forces_and_stress",required:!1,valid_types:"ArrayData",info:"Optional forces and stress"},{name:"ion_files",required:!1,valid_types:"IonData",info:""},{name:"optical_eps2",required:!1,valid_types:"ArrayData",info:"Optional eps2 optical data"},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Optional relaxed structure"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."},{status:403,message:"Basis polarization problem."},{status:404,message:"Error in the parsing of bands"}]},class:"aiida_siesta.workflows.base:SiestaBaseWorkChain"},"siesta.eos":{description:["WorkChain to calculate the equation of state of a solid.",""," The cell shape is fixed, only the volume is rescaled."," In particular the volumes considered are 7 equidistant volumes"," around a starting volume. The starting volume is"," an optional input of the WorkChain (called volume_per_atom)."," If not specified, the input structure volume is used with no modifications."," All the SiestaBaseWorkChain inputs are other inputs of the workchain."," This WorkChain also tries to perform a Birch_Murnaghan fit"," on the calculatad E(V) data."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"batch_size",required:!1,valid_types:"Int",info:`The maximum number of simulations that should run at the same time. + You can set this to a very large number to make sure that all simulations run in + one single batch if you want.`},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"iterate_mode",required:!1,valid_types:"Str",info:`Indicates the way the parameters should be iterated. + Currently allowed values are: + - 'zip': zips all the parameters together (all parameters should + have the same number of values!) + - 'product': performs a cartesian product of the parameters. That is, + all possible combinations of parameters and values are explored. + `},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"scales",required:!1,valid_types:"",info:` + Factors by which the structure should be scaled. + `},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"},{name:"volume_per_atom",required:!1,valid_types:"Float, NoneType",info:"Volume per atom around which to perform the EqOfState"}],outputs:[{name:"results_dict",required:!0,valid_types:"Dict",info:"Containing the calculated E(V) data and, if the fit is sucessfull, the resulting fit parameters"},{name:"equilibrium_structure",required:!1,valid_types:"StructureData",info:"Equilibrium volume structure. Returned only if the fit is succesfull"}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_siesta.workflows.eos:EqOfStateFixedCellShape"},"siesta.bandgap":{description:["Workchain to obtain the bands and bandgap of a structure through Siesta.","",' If "bandskpoints" are set in inputs, it behaves like `SiestaBaseWorkChain`'," adding just the bandgap calculation at the end. If no bandskpoints"," was specified, the bands are computed anyway on a kpoints path automatically"," assigned using seekpath and the input (output) structure"," of the single-point (relaxation/md) calculation."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"seekpath_dict",required:!1,valid_types:"Dict",info:"dictionary of seekpath parameters that are pased to `get_explicit_kpoints_path`"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The calculation results"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"band_gap_info",required:!1,valid_types:"Dict",info:""},{name:"bands",required:!1,valid_types:"BandsData",info:"Optional band structure"},{name:"forces_and_stress",required:!1,valid_types:"ArrayData",info:"Optional forces and stress"},{name:"ion_files",required:!1,valid_types:"IonData",info:""},{name:"optical_eps2",required:!1,valid_types:"ArrayData",info:"Optional eps2 optical data"},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Optional relaxed structure"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"The main SiestaBaseWorkChain failed"},{status:201,message:"The SiestaBaseWorkChain to obtain the bands failed"}]},class:"aiida_siesta.workflows.bandgap:BandgapWorkChain"},"siesta.stm":{description:["STM Workchain.",""," This workchain runs a DFT calculation with siesta, calculates"," the local density of states in an energy window specified by the user (stored"," in a .LDOS file) and post-process it in order to produce simulated STM images."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"emax",required:!0,valid_types:"Float",info:"Higher boundary energy (in eV respect to Ef) for LDOS calculation"},{name:"emin",required:!0,valid_types:"Float",info:"Lower boundary energy (in eV respect to Ef) for LDOS calculation"},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"stm_code",required:!0,valid_types:"Code",info:"STM plstm code"},{name:"stm_mode",required:!0,valid_types:"Str",info:'Allowed values are "constant-height" or "constant-current"'},{name:"stm_spin",required:!0,valid_types:"Str",info:'Allowed values are "none", "collinear" or "non-collinear"'},{name:"stm_value",required:!0,valid_types:"Float",info:"Value of height in Ang or value of current in e/bohr**3"},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"},{name:"stm_options",required:!1,valid_types:"Dict, NoneType",info:"STM plstm code resources and options"}],outputs:[{name:"stm_array",required:!0,valid_types:"ArrayData",info:""},{name:"output_structure",required:!1,valid_types:"StructureData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"The main SiestaBaseWorkChain failed"},{status:201,message:"The SiestaBaseWorkChain to obtain the .LDOS file failed"},{status:202,message:"The STM post-process failed"}]},class:"aiida_siesta.workflows.stm:SiestaSTMWorkChain"},"siesta.baseneb":{description:["Workchain to run a NEB MEP optimization starting from a guessed path.",""," In theory, such task can be accomplished using directly the SiestaCalculation"," and passing the guessed path as xyz files in lua.input_files input (see"," `examples/plugins/siesta/example_neb.py`). Here, instead, the"," guessed path must be specified as a set of structures in a `TrajectoryData` object."," The structures in `TrajectoryData` are then transformed in xyz files and placed"," in a directory that is the passed to lua.input_files when the SiestaCalculation is called."," This better preserves the provenance. Moreover allows cleaner use of ghost (often necessaries)"," Finally, we have a dedicated output containing all the NEB quantities."," This workchain can also become the place where to deal with possible errors due"," to the lua features."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"neb_script",required:!0,valid_types:"SinglefileData",info:"Lua script for NEB engine"},{name:"options",required:!0,valid_types:"Dict",info:"Options"},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"starting_path",required:!0,valid_types:"TrajectoryData",info:"Starting Path"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"},{name:"spring_constant",required:!1,valid_types:"Float",info:""}],outputs:[{name:"neb_output_package",required:!0,valid_types:"TrajectoryData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"The NEB calculation failed"},{status:202,message:"The .xyz files or the NEB.results file could not be retrieved"}]},class:"aiida_siesta.workflows.neb_base:SiestaBaseNEBWorkChain"},"siesta.epsilon":{description:["Workchain to obtain the electronic contribution to the static dielectric constant using Siesta."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Input code"},{name:"optical",required:!0,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters"},{name:"structure",required:!0,valid_types:"StructureData",info:"Input structure"},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The calculation results"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"Optional band structure"},{name:"epsilon",required:!1,valid_types:"Float",info:"Low-frequency dielectric constant"},{name:"forces_and_stress",required:!1,valid_types:"ArrayData",info:"Optional forces and stress"},{name:"ion_files",required:!1,valid_types:"IonData",info:""},{name:"optical_eps2",required:!1,valid_types:"ArrayData",info:"Optional eps2 optical data"},{name:"output_structure",required:!1,valid_types:"StructureData",info:"Optional relaxed structure"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."},{status:403,message:"Basis polarization problem."},{status:404,message:"Error in the parsing of bands"}]},class:"aiida_siesta.workflows.epsilon:EpsilonWorkChain"},"siesta.iterator":{description:["Iterator for the SietaBaseWorkChain.",""," The iterator is extended to iterate over any Siesta keyword."," WARNING: if a keyword not recognized by Siesta is used in `iterate_over`, the iterator will not"," complain. It will just add the keyword to the parameters dict and run the calculation!"],spec:{inputs:[{name:"iterate_over",required:!0,valid_types:"Dict",info:`A dictionary where each key is the name of a parameter we want to iterate + over (str) and each value is a list with all the values to iterate over for + that parameter. Each value in the list can be either a node (unstored or stored) + or a simple python object (str, float, int, bool). + Note that each subclass might parse this keys and values differently, so you should + know how they do it. + `},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"batch_size",required:!1,valid_types:"Int",info:`The maximum number of simulations that should run at the same time. + You can set this to a very large number to make sure that all simulations run in + one single batch if you want.`},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"code",required:!1,valid_types:"Code",info:"Input code"},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"iterate_mode",required:!1,valid_types:"Str",info:`Indicates the way the parameters should be iterated. + Currently allowed values are: + - 'zip': zips all the parameters together (all parameters should + have the same number of values!) + - 'product': performs a cartesian product of the parameters. That is, + all possible combinations of parameters and values are explored. + `},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"options",required:!1,valid_types:"Dict",info:""},{name:"parameters",required:!1,valid_types:"Dict",info:"Input parameters"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"},{name:"structure",required:!1,valid_types:"StructureData",info:"Input structure"}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_siesta.workflows.iterate:SiestaIterator"},"siesta.converger":{description:["Only connects the two parent classes."],spec:{inputs:[{name:"iterate_over",required:!0,valid_types:"Dict",info:`A dictionary where each key is the name of a parameter we want to iterate + over (str) and each value is a list with all the values to iterate over for + that parameter. Each value in the list can be either a node (unstored or stored) + or a simple python object (str, float, int, bool). + Note that each subclass might parse this keys and values differently, so you should + know how they do it. + `},{name:"bandskpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints for bands"},{name:"basis",required:!1,valid_types:"Dict, NoneType",info:"Input basis"},{name:"batch_size",required:!1,valid_types:"Int",info:`The maximum number of simulations that should run at the same time. + You can set this to a very large number to make sure that all simulations run in + one single batch if you want.`},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"code",required:!1,valid_types:"Code",info:"Input code"},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"ions",required:!1,valid_types:"IonData",info:"Input ion file"},{name:"iterate_mode",required:!1,valid_types:"Str",info:`Indicates the way the parameters should be iterated. + Currently allowed values are: + - 'zip': zips all the parameters together (all parameters should + have the same number of values!) + - 'product': performs a cartesian product of the parameters. That is, + all possible combinations of parameters and values are explored. + `},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Input kpoints"},{name:"lua",required:!1,valid_types:"",info:"Script and files for the Lua engine"},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"optical",required:!1,valid_types:"Dict, NoneType",info:"Specifications for optical properties"},{name:"options",required:!1,valid_types:"Dict",info:""},{name:"parameters",required:!1,valid_types:"Dict",info:"Input parameters"},{name:"parent_calc_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Parent folder"},{name:"pseudo_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"pseudos",required:!1,valid_types:"PsfData, PsmlData",info:"Input pseudo potentials"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Input settings"},{name:"structure",required:!1,valid_types:"StructureData",info:"Input structure"},{name:"target",required:!1,valid_types:"Str, NoneType",info:"The parameter that you want to track."},{name:"threshold",required:!1,valid_types:"Int, Float",info:"The maximum difference between two consecutive steps to consider that convergence is reached"}],outputs:[{name:"converged",required:!0,valid_types:"",info:"Whether the target has converged"},{name:"converged_parameters",required:!1,valid_types:"",info:"The values for the parameters that was enough to achieve convergence. If converged is not achieved, it won't be returned"},{name:"converged_target_value",required:!1,valid_types:"",info:"The value of the target with convergence reached."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_siesta.workflows.converge:SiestaConverger"},"siesta.sequential_converger":{description:["An iterator of convergers."],spec:{inputs:[{name:"converger_inputs",required:!0,valid_types:"Data",info:""},{name:"iterate_over",required:!0,valid_types:"List",info:`A dictionary where each key is the name of a parameter we want to iterate + over (str) and each value is a list with all the values to iterate over for + that parameter. Each value in the list can be either a node (unstored or stored) + or a simple python object (str, float, int, bool). + Note that each subclass might parse this keys and values differently, so you should + know how they do it. + `},{name:"batch_size",required:!1,valid_types:"Int",info:`The maximum number of simulations that should run at the same time. + You can set this to a very large number to make sure that all simulations run in + one single batch if you want.`},{name:"iterate_mode",required:!1,valid_types:"Str",info:`Indicates the way the parameters should be iterated. + Currently allowed values are: + - 'zip': zips all the parameters together (all parameters should + have the same number of values!) + - 'product': performs a cartesian product of the parameters. That is, + all possible combinations of parameters and values are explored. + `},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"converged_parameters",required:!0,valid_types:"",info:"The values for the parameters that was enough to achieve convergence. If convergence is not achieved, it will be an empty dictionary"},{name:"converged_target_value",required:!1,valid_types:"",info:"The value of the target with convergence reached."},{name:"unconverged_parameters",required:!1,valid_types:"",info:"The list of unconverged parameters."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_siesta.workflows.converge:SiestaSequentialConverger"},"siesta.forbasisopt":{description:["Wrapper of SiestaBaseWorkChain.",""," Class wrapping the SiestaBaseWorkChain with the scope of"," assigning the variables of the optimization process to the right"," input of the SiestaBaseWorkChain."],spec:{inputs:[{name:"lower_bounds",required:!0,valid_types:"List",info:""},{name:"out_name",required:!0,valid_types:"Str",info:""},{name:"siesta_base",required:!0,valid_types:"Data",info:""},{name:"the_names",required:!0,valid_types:"List",info:""},{name:"the_values",required:!0,valid_types:"List",info:""},{name:"upper_bounds",required:!0,valid_types:"List",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"ene",required:!0,valid_types:"Float",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_siesta.workflows._for_optimization:ForBasisOptWorkChain"},"siesta.simplex_basis":{description:["Workchain running a simple NelderMead optimization (simplex) varing variables defined in the basis dictionaries."],spec:{inputs:[{name:"siesta_base",required:!0,valid_types:"Data",info:""},{name:"simplex",required:!0,valid_types:"",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"last_simplex",required:!0,valid_types:"List",info:""},{name:"optimal_process_output",required:!0,valid_types:"",info:"Output value of the optimal evaluation process."},{name:"optimal_process_uuid",required:!0,valid_types:"",info:"UUID of the optimal evaluation process."},{name:"optimal_process_input",required:!1,valid_types:"",info:"Input value of the optimal evaluation process."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"The OptimizationWorkChain did not finished"}]},class:"aiida_siesta.workflows.simplex_basis:SimplexBasisOptimization"},"siesta.two_step_basis_opt":{description:["Optimization that is more similar to the simplex code in the siesta utils.","",' The optimization has two levels, a "marcrostep" that consists in the restart of'," a simplex with gradual reduction of the dimention of the initial simplex."],spec:{inputs:[{name:"macrostep",required:!0,valid_types:"",info:""},{name:"siesta_base",required:!0,valid_types:"Data",info:""},{name:"simplex",required:!0,valid_types:"",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"last_simplex",required:!0,valid_types:"List",info:""},{name:"optimal_process_output",required:!0,valid_types:"",info:"Output value of the optimal evaluation process."},{name:"optimal_process_uuid",required:!0,valid_types:"",info:"UUID of the optimal evaluation process."},{name:"optimal_process_input",required:!1,valid_types:"",info:"Input value of the optimal evaluation process."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_siesta.workflows.two_steps_optimization:TwoStepsBasisOpt"},"siesta.basis_optimization":{description:["WorkChain for basis optimization."],spec:{inputs:[{name:"optimization_schema",required:!0,valid_types:"",info:""},{name:"siesta_base",required:!0,valid_types:"Data",info:""},{name:"simplex",required:!0,valid_types:"",info:""},{name:"add_orbital",required:!1,valid_types:"Dict, NoneType",info:""},{name:"basis_sizes",required:!1,valid_types:"List",info:""},{name:"metadata",required:!1,valid_types:"",info:""},{name:"sizes_monitored_quantity",required:!1,valid_types:"Str",info:""}],outputs:[{name:"optimal_basis_block",required:!0,valid_types:"Dict",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:200,message:"The SiestaIterator running pao sizes failed"},{status:201,message:"The basis optimization failed, probably not sufficient steps"},{status:202,message:"The basis optimization failed, probably not sufficient steps"}]},class:"aiida_siesta.workflows.basis_optimization:BasisOptimizationWorkChain"}},"aiida.data":{"siesta.ion":"aiida_siesta.data.ion:IonData"},"aiida.tools.data.orbitals":{"siesta.atomic_orbital":"aiida_siesta.data.atomic_orbitals:SislAtomicOrbital"}},commits_count:0,development_status:"stable",errors:[],warnings:["W019: No bdist_wheel available for PyPI release."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:13},{colorclass:"orange",text:"Other (Tools data orbitals)",count:1}],pip_install_cmd:"pip install aiida-siesta",is_installable:"True"},"aiida-spex":{code_home:"https://github.com/JuDFTteam/aiida-spex",entry_point_prefix:"spex",pip_url:"git+https://github.com/JuDFTteam/aiida-spex",name:"aiida-spex",package_name:"aiida_spex",hosted_on:"github.com",metadata:{author:"The SPEX Team",author_email:"a.chandran@fz-juelich.de",version:"1.1.2",description:"AiiDA plugin for SPEX code",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=1.0.0b3,<3.0.0",entry_points:{"aiida.calculations":{"spex.spex":"aiida_spex.calculations.spex:SpexCalculation"},"aiida.data":{"spex.spexinp":"aiida_spex.data.spexinp:SpexinpData"},"aiida.parsers":{"spex.spexparser":"aiida_spex.parsers.spex:SpexParser"},"aiida.workflows":{"spex.job":"aiida_spex.workflows.job:SpexJobWorkchain"}},commits_count:0,development_status:"planning",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install git+https://github.com/JuDFTteam/aiida-spex"},"aiida-spirit":{code_home:"https://github.com/JuDFTteam/aiida-spirit/tree/main",documentation_url:"https://aiida-spirit.readthedocs.io/",entry_point_prefix:"spirit",name:"aiida-spirit",pip_url:"aiida-spirit",package_name:"aiida_spirit",hosted_on:"github.com",metadata:{release_date:"2023-06-23",description:"AiiDA plugin for the spirit code",author:"The JuDFT Team",author_email:"p.ruessmann@fz-juelich.de",license:"MIT",home_page:"https://github.com/JuDFTteam/aiida-spirit",classifiers:["Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.2.2"},aiida_version:null,entry_points:{"aiida.calculations":{spirit:"aiida_spirit.calculations:SpiritCalculation"},"aiida.parsers":{spirit:"aiida_spirit.parsers:SpiritParser"}},commits_count:0,development_status:"planning",errors:[],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install aiida-spirit"},"aiida-ssh2win":{entry_point_prefix:"ssh2win",code_home:"https://github.com/edan-bainglass/aiida-ssh2win",version_file:"https://raw.githubusercontent.com/edan-bainglass/aiida-ssh2win/develop/aiida_ssh2win/__init__.py",pip_url:"git+https://github.com/edan-bainglass/aiida-ssh2win",name:"aiida-ssh2win",package_name:"aiida_ssh2win",hosted_on:"github.com",metadata:{},aiida_version:null,entry_points:{},commits_count:0,development_status:"planning",errors:[],warnings:["E001: Cannot fetch all data from PyPI and missing plugin_info key!","W002: AiiDA version not found"],summaryinfo:[],pip_install_cmd:"pip install git+https://github.com/edan-bainglass/aiida-ssh2win"},"aiida-sshonly":{code_home:"https://github.com/adegomme/aiida-sshonly",development_status:"beta",entry_point_prefix:"sshonly",pip_url:"aiida-sshonly",plugin_info:"https://raw.github.com/adegomme/aiida-sshonly/master/setup.json",name:"aiida-sshonly",package_name:"aiida_sshonly",hosted_on:"github.com",metadata:{release_date:"2020-10-07",description:"AiiDA plugin adding a sshonly transport option, using only SSH to transfer files, avoiding SFTP, in case it's blocked or non functional on a remote system",author:"adegomme",license:"MIT",home_page:"https://github.com/adegomme/aiida-sshonly",classifiers:["Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.1.0a3"},aiida_version:">=1.3.0,<2.0.0",entry_points:{"aiida.transports":{ssh_only:"aiida_sshonly.transports.sshonly:SshOnlyTransport"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-sshonly
Collecting aiida-sshonly
+  Downloading aiida-sshonly-0.1.0a3.tar.gz (19 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-sshonly to determine which version is compatible with other requirements. This could take a while.
+  Downloading aiida_sshonly-0.1.0a2-py3-none-any.whl.metadata (3.4 kB)
+  Downloading aiida_sshonly-0.1.0a1-py3-none-any.whl.metadata (3.4 kB)
+  Downloading aiida-sshonly-0.1.0a0.tar.gz (19 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+ERROR: Cannot install aiida-sshonly==0.1.0a0, aiida-sshonly==0.1.0a1, aiida-sshonly==0.1.0a2 and aiida-sshonly==0.1.0a3 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-sshonly 0.1.0a3 depends on aiida-core<2.0.0 and >=1.3.0
+    aiida-sshonly 0.1.0a2 depends on aiida-core<2.0.0 and >=1.3.0
+    aiida-sshonly 0.1.0a1 depends on aiida-core<2.0.0 and >=1.3.0
+    aiida-sshonly 0.1.0a0 depends on aiida-core<2.0.0 and >=1.1.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.","W010: Entry point 'ssh_only' does not start with prefix 'sshonly.'"],summaryinfo:[{colorclass:"orange",text:"Other (Transports)",count:1}],pip_install_cmd:"pip install --pre aiida-sshonly",is_installable:"False"},"aiida-statefile-schedulers":{code_home:"https://github.com/dev-zero/aiida-statefile-schedulers",development_status:"beta",entry_point_prefix:"statefile_schedulers",pip_url:"aiida-statefile-schedulers",name:"aiida-statefile-schedulers",package_name:"aiida_statefile_schedulers",hosted_on:"github.com",metadata:{release_date:"2021-11-23",description:"Simple statefile-driven task schedulers for AiiDA",author:"Tiziano Müller",author_email:"tm@dev-zero.ch",license:"MIT",home_page:"https://github.com/dev-zero/aiida-statefile-schedulers",classifiers:["Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.2.1"},aiida_version:null,entry_points:{"aiida.schedulers":{"statefile_schedulers.direct":"aiida_statefile_schedulers.schedulers.direct:StatefileDirectScheduler"}},commits_count:0,errors:[],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"orange",text:"Other (Schedulers)",count:1}],pip_install_cmd:"pip install aiida-statefile-schedulers",is_installable:"True"},"aiida-strain":{code_home:"https://github.com/greschd/aiida-strain",documentation_url:"https://aiida-strain.readthedocs.io",entry_point_prefix:"strain",pip_url:"aiida-strain",name:"aiida-strain",package_name:"aiida_strain",hosted_on:"github.com",metadata:{release_date:"2019-11-22",description:"AiiDA Plugin for applying strain to structures",author:"Dominik Gresch",author_email:"greschd@gmx.ch",license:"Apache 2.0",home_page:"https://aiida-strain.readthedocs.io",classifiers:["Development Status :: 3 - Alpha","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: Apache Software License","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Scientific/Engineering :: Physics"],version:"0.2.0"},aiida_version:null,entry_points:{"aiida.workflows":{"strain.apply_strains":{description:["Workchain to create strained structures from a given input structure."],spec:{inputs:[{name:"strain_kind",required:!0,valid_types:"Str",info:""},{name:"strain_parameters",required:!0,valid_types:"Str",info:""},{name:"strain_strengths",required:!0,valid_types:"List",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_strain:ApplyStrains"},"strain.apply_strains_with_symmetry":{description:["Workchain to create strained structures from an input structure, and select the symmetries which are compatible with the strained structure from a set of given input symmetries."],spec:{inputs:[{name:"strain_kind",required:!0,valid_types:"Str",info:""},{name:"strain_parameters",required:!0,valid_types:"Str",info:""},{name:"strain_strengths",required:!0,valid_types:"List",info:""},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"symmetries",required:!0,valid_types:"SinglefileData",info:""},{name:"symmetry_repr_code",required:!0,valid_types:"Code",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_strain:ApplyStrainsWithSymmetry"}}},commits_count:0,development_status:"alpha",errors:[],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"green",text:"Workflows",count:2}],pip_install_cmd:"pip install aiida-strain",is_installable:"True"},"aiida-supercell":{code_home:"https://github.com/pzarabadip/aiida-supercell",development_status:"stable",documentation_url:"https://aiida-supercell.readthedocs.io/",entry_point_prefix:"supercell",pip_url:"git+https://github.com/pzarabadip/aiida-supercell",name:"aiida-supercell",package_name:"aiida_supercell",hosted_on:"github.com",metadata:{author:"Pezhman Zarabadi-Poor",author_email:"pzarabadip@gmail.com",version:"1.0.1",description:"AiiDA Plugin for Supercell program",classifiers:["Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"]},aiida_version:">=1.0.0,<2.0",entry_points:{"aiida.calculations":{supercell:"aiida_supercell.calculations:SupercellCalculation"},"aiida.parsers":{supercell:"aiida_supercell.parsers:SupercellParser"}},commits_count:0,errors:[`E001: Failed to install plugin aiida-supercell
Collecting git+https://github.com/pzarabadip/aiida-supercell
+  Cloning https://github.com/pzarabadip/aiida-supercell to /tmp/pip-req-build-4r96dutg
+  Running command git clone --filter=blob:none --quiet https://github.com/pzarabadip/aiida-supercell /tmp/pip-req-build-4r96dutg
+  Resolved https://github.com/pzarabadip/aiida-supercell to commit 24a89f990fcf84584a51f49b53311a5c6db18347
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'done'
+INFO: pip is looking at multiple versions of aiida-supercell to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-supercell==1.0.1 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-supercell 1.0.1 depends on aiida-core<2.0 and >=1.0.0
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install git+https://github.com/pzarabadip/aiida-supercell",is_installable:"False"},"aiida-symmetry-representation":{code_home:"https://github.com/greschd/aiida_symmetry_representation",documentation_url:"https://aiida-symmetry-representation.readthedocs.io",entry_point_prefix:"symmetry_representation",pip_url:"aiida-symmetry-representation",name:"aiida-symmetry-representation",package_name:"aiida_symmetry_representation",hosted_on:"github.com",metadata:{release_date:"2019-11-18",description:"AiiDA Plugin for symmetry representations.",author:"Dominik Gresch",author_email:"greschd@gmx.ch",license:"Apache 2.0",home_page:"https://aiida-symmetry-representation.readthedocs.io",classifiers:["Development Status :: 5 - Production/Stable","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: Apache Software License","Programming Language :: Python :: 3","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Topic :: Scientific/Engineering :: Physics"],version:"0.2.0"},aiida_version:null,entry_points:{"aiida.calculations":{"symmetry_representation.filter_symmetries":{description:["Calculation class to run the ``symmetry-repr filter_symmetries`` command."],spec:{inputs:[{name:"structure",required:!0,valid_types:"StructureData",info:"Structure with which the filtered symmetries should be compatible."},{name:"symmetries",required:!0,valid_types:"SinglefileData",info:"File containing the symmetries (in HDF5 format)."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"symmetries",required:!0,valid_types:"SinglefileData",info:"The HDF5 file containing the symmetries which are compatible with the structure."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_symmetry_representation.calculations.filter_symmetries:FilterSymmetriesCalculation"}},"aiida.parsers":{"symmetry_representation.symmetry":"aiida_symmetry_representation.parsers.symmetries:SymmetriesParser"}},commits_count:0,development_status:"stable",errors:[],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install aiida-symmetry-representation",is_installable:"True"},"aiida-tbextraction":{code_home:"https://github.com/greschd/aiida-tbextraction",documentation_url:"https://aiida-tbextraction.readthedocs.io/",entry_point_prefix:"tbextraction",pip_url:"aiida-tbextraction",name:"aiida-tbextraction",package_name:"aiida_tbextraction",hosted_on:"github.com",metadata:{release_date:"2020-02-25",description:"AiiDA Plugin for extracting tight-binding models",author:"Dominik Gresch",author_email:"greschd@gmx.ch",license:"Apache 2.0",home_page:"https://aiida-tbextraction.readthedocs.io",classifiers:["Development Status :: 4 - Beta","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: Apache Software License","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Scientific/Engineering :: Physics"],version:"0.2.0b1"},aiida_version:null,entry_points:{"aiida.workflows":{"tbextraction.fp_run.base":"aiida_tbextraction.fp_run:FirstPrinciplesRunBase","tbextraction.fp_run.reference_bands.base":"aiida_tbextraction.fp_run.reference_bands:ReferenceBandsBase","tbextraction.fp_run.wannier_input.base":"aiida_tbextraction.fp_run.wannier_input:WannierInputBase","tbextraction.calculate_tb":"aiida_tbextraction.calculate_tb:TightBindingCalculation","tbextraction.model_evaluation.base":"aiida_tbextraction.model_evaluation:ModelEvaluationBase","tbextraction.model_evaluation.band_difference":"aiida_tbextraction.model_evaluation:BandDifferenceModelEvaluation","tbextraction.energy_windows.run_window":"aiida_tbextraction.energy_windows.run_window:RunWindow","tbextraction.energy_windows.window_search":"aiida_tbextraction.energy_windows.window_search:WindowSearch","tbextraction.optimize_fp_tb":"aiida_tbextraction.optimize_fp_tb:OptimizeFirstPrinciplesTightBinding","tbextraction.optimize_strained_fp_tb":"aiida_tbextraction.optimize_strained_fp_tb:OptimizeStrainedFirstPrinciplesTightBinding"}},commits_count:0,development_status:"beta",errors:[`E001: Failed to install plugin aiida-tbextraction
Collecting aiida-tbextraction
+  Downloading aiida-tbextraction-0.2.0b1.tar.gz (19 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'error'
+  error: subprocess-exited-with-error
+  
+  × python setup.py egg_info did not run successfully.
+  │ exit code: 1
+  ╰─> [20 lines of output]
+      /opt/conda/lib/python3.10/site-packages/setuptools/__init__.py:80: _DeprecatedInstaller: setuptools.installer and fetch_build_eggs are deprecated.
+      !!
+      
+              ********************************************************************************
+              Requirements should be satisfied by a PEP 517 installer.
+              If you are using pip, you can try \`pip install --use-pep517\`.
+              ********************************************************************************
+      
+      !!
+        dist.fetch_build_eggs(dist.setup_requires)
+      [ REENTRY ] registering entry points with reentry...
+      [ REENTRY ] ... registered to /home/aiida/.config/reentry/data/5d9e5327e684dbfa000493a11d0110f0abb5242e20a9b88369b9910dfe0d9564
+      [ REENTRY ] Following entrypoints were registered
+      
+          aiida-tbextraction -> {'aiida.workflows': {'tbextraction.fp_run.base': 'tbextraction.fp_run.base = aiida_tbextraction.fp_run:FirstPrinciplesRunBase', 'tbextraction.fp_run.reference_bands.base': 'tbextraction.fp_run.reference_bands.base = aiida_tbextraction.fp_run.reference_bands:ReferenceBandsBase', 'tbextraction.fp_run.wannier_input.base': 'tbextraction.fp_run.wannier_input.base = aiida_tbextraction.fp_run.wannier_input:WannierInputBase', 'tbextraction.calculate_tb': 'tbextraction.calculate_tb = aiida_tbextraction.calculate_tb:TightBindingCalculation', 'tbextraction.model_evaluation.base': 'tbextraction.model_evaluation.base = aiida_tbextraction.model_evaluation:ModelEvaluationBase', 'tbextraction.model_evaluation.band_difference': 'tbextraction.model_evaluation.band_difference = aiida_tbextraction.model_evaluation:BandDifferenceModelEvaluation', 'tbextraction.energy_windows.run_window': 'tbextraction.energy_windows.run_window = aiida_tbextraction.energy_windows.run_window:RunWindow', 'tbextraction.energy_windows.window_search': 'tbextraction.energy_windows.window_search = aiida_tbextraction.energy_windows.window_search:WindowSearch', 'tbextraction.optimize_fp_tb': 'tbextraction.optimize_fp_tb = aiida_tbextraction.optimize_fp_tb:OptimizeFirstPrinciplesTightBinding', 'tbextraction.optimize_strained_fp_tb': 'tbextraction.optimize_strained_fp_tb = aiida_tbextraction.optimize_strained_fp_tb:OptimizeStrainedFirstPrinciplesTightBinding'}}
+      [ REENTRY ] Current entry point map at /home/aiida/.config/reentry/data/5d9e5327e684dbfa000493a11d0110f0abb5242e20a9b88369b9910dfe0d9564:
+          aiida-tbextraction -> {'aiida.workflows': {'tbextraction.fp_run.base': 'tbextraction.fp_run.base = aiida_tbextraction.fp_run:FirstPrinciplesRunBase', 'tbextraction.fp_run.reference_bands.base': 'tbextraction.fp_run.reference_bands.base = aiida_tbextraction.fp_run.reference_bands:ReferenceBandsBase', 'tbextraction.fp_run.wannier_input.base': 'tbextraction.fp_run.wannier_input.base = aiida_tbextraction.fp_run.wannier_input:WannierInputBase', 'tbextraction.calculate_tb': 'tbextraction.calculate_tb = aiida_tbextraction.calculate_tb:TightBindingCalculation', 'tbextraction.model_evaluation.base': 'tbextraction.model_evaluation.base = aiida_tbextraction.model_evaluation:ModelEvaluationBase', 'tbextraction.model_evaluation.band_difference': 'tbextraction.model_evaluation.band_difference = aiida_tbextraction.model_evaluation:BandDifferenceModelEvaluation', 'tbextraction.energy_windows.run_window': 'tbextraction.energy_windows.run_window = aiida_tbextraction.energy_windows.run_window:RunWindow', 'tbextraction.energy_windows.window_search': 'tbextraction.energy_windows.window_search = aiida_tbextraction.energy_windows.window_search:WindowSearch', 'tbextraction.optimize_fp_tb': 'tbextraction.optimize_fp_tb = aiida_tbextraction.optimize_fp_tb:OptimizeFirstPrinciplesTightBinding', 'tbextraction.optimize_strained_fp_tb': 'tbextraction.optimize_strained_fp_tb = aiida_tbextraction.optimize_strained_fp_tb:OptimizeStrainedFirstPrinciplesTightBinding'}}
+      error in aiida-tbextraction setup command: 'install_requires' must be a string or list of strings containing valid project/version requirement specifiers; Expected end or semicolon (after version specifier)
+          aiida-core>=1.0.0<2
+                    ~~~~~~~^
+      [end of output]
+  
+  note: This error originates from a subprocess, and is likely not a problem with pip.
+error: metadata-generation-failed
+
+× Encountered error while generating package metadata.
+╰─> See above for output.
+
+note: This is an issue with the package mentioned above, not pip.
+hint: See above for details.
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"green",text:"Workflows",count:10}],pip_install_cmd:"pip install --pre aiida-tbextraction",is_installable:"False"},"aiida-tbmodels":{code_home:"https://github.com/greschd/aiida-tbmodels",documentation_url:"https://aiida-tbmodels.readthedocs.io",entry_point_prefix:"tbmodels",pip_url:"aiida-tbmodels",name:"aiida-tbmodels",package_name:"aiida_tbmodels",hosted_on:"github.com",metadata:{release_date:"2020-03-03",description:"AiiDA Plugin for running TBmodels",author:"Dominik Gresch",author_email:"greschd@gmx.ch",license:"Apache 2.0",home_page:"https://aiida-tbmodels.readthedocs.io",classifiers:["Development Status :: 3 - Alpha","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: Apache Software License","Programming Language :: Python :: 3","Programming Language :: Python :: 3.6","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Topic :: Scientific/Engineering :: Physics"],version:"0.3.0"},aiida_version:null,entry_points:{"aiida.calculations":{"tbmodels.eigenvals":"aiida_tbmodels.calculations.eigenvals:EigenvalsCalculation","tbmodels.parse":"aiida_tbmodels.calculations.parse:ParseCalculation","tbmodels.slice":"aiida_tbmodels.calculations.slice:SliceCalculation","tbmodels.symmetrize":"aiida_tbmodels.calculations.symmetrize:SymmetrizeCalculation"},"aiida.parsers":{"tbmodels.model":"aiida_tbmodels.parsers.model:ModelParser"}},commits_count:0,development_status:"alpha",errors:[`E001: Failed to install plugin aiida-tbmodels
Collecting aiida-tbmodels
+  Downloading aiida-tbmodels-0.3.0.tar.gz (8.8 kB)
+  Preparing metadata (setup.py): started
+  Preparing metadata (setup.py): finished with status 'error'
+  error: subprocess-exited-with-error
+  
+  × python setup.py egg_info did not run successfully.
+  │ exit code: 1
+  ╰─> [20 lines of output]
+      /opt/conda/lib/python3.10/site-packages/setuptools/__init__.py:80: _DeprecatedInstaller: setuptools.installer and fetch_build_eggs are deprecated.
+      !!
+      
+              ********************************************************************************
+              Requirements should be satisfied by a PEP 517 installer.
+              If you are using pip, you can try \`pip install --use-pep517\`.
+              ********************************************************************************
+      
+      !!
+        dist.fetch_build_eggs(dist.setup_requires)
+      [ REENTRY ] registering entry points with reentry...
+      [ REENTRY ] ... registered to /home/aiida/.config/reentry/data/5d9e5327e684dbfa000493a11d0110f0abb5242e20a9b88369b9910dfe0d9564
+      [ REENTRY ] Following entrypoints were registered
+      
+          aiida-tbmodels -> {'aiida.calculations': {'tbmodels.eigenvals': 'tbmodels.eigenvals = aiida_tbmodels.calculations.eigenvals:EigenvalsCalculation', 'tbmodels.parse': 'tbmodels.parse = aiida_tbmodels.calculations.parse:ParseCalculation', 'tbmodels.slice': 'tbmodels.slice = aiida_tbmodels.calculations.slice:SliceCalculation', 'tbmodels.symmetrize': 'tbmodels.symmetrize = aiida_tbmodels.calculations.symmetrize:SymmetrizeCalculation'}, 'aiida.parsers': {'tbmodels.model': 'tbmodels.model = aiida_tbmodels.parsers.model:ModelParser'}}
+      [ REENTRY ] Current entry point map at /home/aiida/.config/reentry/data/5d9e5327e684dbfa000493a11d0110f0abb5242e20a9b88369b9910dfe0d9564:
+          aiida-tbmodels -> {'aiida.calculations': {'tbmodels.eigenvals': 'tbmodels.eigenvals = aiida_tbmodels.calculations.eigenvals:EigenvalsCalculation', 'tbmodels.parse': 'tbmodels.parse = aiida_tbmodels.calculations.parse:ParseCalculation', 'tbmodels.slice': 'tbmodels.slice = aiida_tbmodels.calculations.slice:SliceCalculation', 'tbmodels.symmetrize': 'tbmodels.symmetrize = aiida_tbmodels.calculations.symmetrize:SymmetrizeCalculation'}, 'aiida.parsers': {'tbmodels.model': 'tbmodels.model = aiida_tbmodels.parsers.model:ModelParser'}}
+      error in aiida-tbmodels setup command: 'install_requires' must be a string or list of strings containing valid project/version requirement specifiers; Expected end or semicolon (after version specifier)
+          aiida-core>=1.1.0<2.0.0
+                    ~~~~~~~^
+      [end of output]
+  
+  note: This error originates from a subprocess, and is likely not a problem with pip.
+error: metadata-generation-failed
+
+× Encountered error while generating package metadata.
+╰─> See above for output.
+
+note: This is an issue with the package mentioned above, not pip.
+hint: See above for details.
+
`],warnings:["W019: No bdist_wheel available for PyPI release.","W002: AiiDA version not found"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:4},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install aiida-tbmodels",is_installable:"False"},"aiida-tcod":{code_home:"https://github.com/aiidateam/aiida-tcod",development_status:"beta",entry_point_prefix:"tcod",pip_url:"git+https://github.com/aiidateam/aiida-tcod",name:"aiida-tcod",package_name:"aiida_tcod",hosted_on:"github.com",metadata:{author:"The AiiDA team",author_email:"developers@aiida.net",version:"0.1.0a0",description:"AiiDA plugin to interact with the TCOD",classifiers:["Programming Language :: Python"]},aiida_version:">=1.0.0b1",entry_points:{"aiida.tools.dbexporters":{tcod:"aiida.tools.dbexporters.tcod"}},commits_count:0,errors:[],warnings:["W003: Missing classifier 'Framework :: AiiDA'","W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"orange",text:"Other (Database Exporters)",count:1}],pip_install_cmd:"pip install git+https://github.com/aiidateam/aiida-tcod",is_installable:"True"},"aiida-uppasd":{code_home:"https://github.com/uppasd/aiida-uppasd",documentation_url:"https://github.com/uppasd/aiida-uppasd/blob/master/README.md",entry_point_prefix:"uppasd",pip_url:"git+https://github.com/unkcpz/aiida-uppasd",name:"aiida-uppasd",package_name:"aiida_uppasd",hosted_on:"github.com",metadata:{author:"Qichen Xu, Anders Bergman, Anna Delin, Jonathan Chico",author_email:"qichenx@kth.se",version:"0.1.0",description:"Interface for UppASD and AiiDA",classifiers:["Programming Language :: Python","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Framework :: AiiDA"]},aiida_version:">=1.1.0,<2.0.0",entry_points:{"aiida.calculations":{UppASD_core_calculations:"UppASD_AiiDA.calculations.core_calcs:UppASD"},"aiida.parsers":{UppASD_core_parsers:"UppASD_AiiDA.parsers.core_parser:SpinDynamic_core_parser"}},commits_count:0,development_status:"planning",errors:[],warnings:["W010: Entry point 'UppASD_core_calculations' does not start with prefix 'uppasd.'","W010: Entry point 'UppASD_core_parsers' does not start with prefix 'uppasd.'"],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install git+https://github.com/unkcpz/aiida-uppasd"},"aiida-vasp":{code_home:"https://github.com/aiida-vasp/aiida-vasp",documentation_url:"https://aiida-vasp.readthedocs.io/",entry_point_prefix:"vasp",pip_url:"aiida-vasp",plugin_info:"https://raw.githubusercontent.com/aiida-vasp/aiida-vasp/master/setup.json",name:"aiida-vasp",package_name:"aiida_vasp",hosted_on:"github.com",metadata:{release_date:"2024-04-23",description:"AiiDA plugin for running VASP calculations and workflows.",author_email:"Espen Flage-Larsen ",classifiers:["Development Status :: 5 - Production/Stable","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Physics"],version:"3.1.1"},aiida_version:">=2.4,<3.0",entry_points:{"aiida.calculations":{"vasp.immigrant":{description:["Parse VASP output objects stored in a specified directory.",""," Simulate running the VaspCalculation up to the point where it can be"," retrieved and parsed, then hand over control to the runner for the rest.",""," Usage examples"," --------------"," Immigrant calculation can be perfomed as follows.",""," ::",""," code = Code.get_from_string('vasp@local')"," folder = '/home/username/vasp-calc-dir'"," settings = {'parser_settings': {'add_energies': True,"," 'add_forces': True,"," 'electronic_step_energies': True}}"," VaspImmigrant = CalculationFactory('vasp.immigrant')"," builder = VaspImmigrant.get_builder_from_folder(code,"," folder,"," settings=settings)"," submit(builder)",""," Instead of ``builder``, inputs dict is obtained similarly as",""," ::",""," code = Code.get_from_string('vasp@local')"," folder = '/home/username/vasp-calc-dir'"," settings = {'parser_settings': {'add_energies': True,"," 'add_forces': True,"," 'electronic_step_energies': True}}"," VaspImmigrant = CalculationFactory('vasp.immigrant')"," inputs = VaspImmigrant.get_inputs_from_folder(code,"," folder,"," settings=settings)"," submit(VaspImmigrant, **inputs)",""," Note"," ----"," The defaul metadata is set automatically as follows::",""," {'options': {'max_wallclock_seconds': 1,"," 'resources': {'num_machines': 1, 'num_mpiprocs_per_machine': 1}}}",""," Specific scheduler may require setting ``resources`` differently"," (e.g., sge ``'parallel_env'``).",""," ``get_inputs_from_folder`` and ``get_builder_from_folder`` accept several"," kwargs, see the docstring of ``get_inputs_from_folder``."],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The kpoints to use (KPOINTS)."},{name:"parameters",required:!0,valid_types:"Dict",info:"The VASP input parameters (INCAR)."},{name:"potential",required:!0,valid_types:"PotcarData",info:"The potentials (POTCAR)."},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR)."},{name:"charge_density",required:!1,valid_types:"ChargedensityData, NoneType",info:"The charge density. (CHGCAR)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"dynamics",required:!1,valid_types:"Dict, NoneType",info:"The VASP parameters related to ionic dynamics, e.g. flags to set the selective dynamics"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"remote_workdir",required:!1,valid_types:"str, NoneType",info:""},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:"A remote folder to restart from if need be"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters not related to VASP itself."},{name:"wavefunctions",required:!1,valid_types:"WavefunData, NoneType",info:"The wave function coefficients. (WAVECAR)"}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:333,message:"VASP did not produce any output and did likely not execute properly."},{status:350,message:"the retrieved folder data node could not be accessed."},{status:351,message:"the retrieved_temporary folder data node could not be accessed."},{status:352,message:"an object that is marked by the parser as critical is missing."},{status:700,message:"Calculation did not reach the end of execution."},{status:701,message:"The electronic structure is not converged."},{status:702,message:"The ionic relaxation is not converged."},{status:703,message:"VASP calculation encountered a critical error: {error_message}."},{status:704,message:"Outputs for diagnosis are missing, please make sure `run_status` and `notifications` quantities are requested for parsing."},{status:1001,message:"parsing an object has failed."},{status:1002,message:"the parser is not able to parse the {quantity} quantity"},{status:1003,message:"the vasprun.xml was truncated and recovery parsing failed to parse at least one of the requested quantities: {quantities}, very likely the VASP calculation did not run properly"},{status:1004,message:"the parser is not able to compose one or more output nodes: {nodes}"},{status:1005,message:"Overflow detected in XML while parsing."}]},class:"aiida_vasp.calcs.immigrant:VaspImmigrant"},"vasp.neb":{description:["NEB calculations using VASP",""," ------------------------------------"," Calculations for performing NEB calculations."," NEB calculations requires standard VASP inputs, but POSCAR are placed in"," folder names 00, 01, 02... N for N-1 number of images.",""," Input frames should be placed under the ``neb_images`` input namespace as a dictionary like::"," {"," 'image_00': structure_1,"," 'image_01': structure_2"," ...."," }",""," Output of individual frames are placed in the corresponding namespace under the same convention."],spec:{inputs:[{name:"final_structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR) for the final image."},{name:"initial_structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR) for initial image."},{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The kpoints to use (KPOINTS)."},{name:"neb_images",required:!0,valid_types:"StructureData, CifData",info:"Starting structure for the NEB images"},{name:"parameters",required:!0,valid_types:"Dict",info:"The VASP input parameters (INCAR)."},{name:"potential",required:!0,valid_types:"PotcarData",info:"The potentials (POTCAR)."},{name:"charge_density",required:!1,valid_types:"ChargedensityData",info:"The charge density. (CHGCAR)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"dynamics",required:!1,valid_types:"Dict, NoneType",info:"The VASP parameters related to ionic dynamics, e.g. flags to set the selective dynamics"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:"A remote folder to restart from if need be"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters not related to VASP itself."},{name:"wavefunctions",required:!1,valid_types:"WavefunData",info:"The wave function coefficients. (WAVECAR)"}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"Per-image misc output."},{name:"neb_misc",required:!0,valid_types:"Dict",info:"NEB related data combined for each image"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"structure",required:!0,valid_types:"StructureData",info:"NEB images"},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"Kpoints for each image."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization for each image."},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output file containing the plane wave coefficients."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:333,message:"VASP did not produce any output files and did likely not execute properly."},{status:350,message:"the retrieved folder data node could not be accessed."},{status:351,message:"the retrieved_temporary folder data node could not be accessed."},{status:352,message:"an object that is marked by the parser as critical is missing."},{status:352,message:"a file that is marked by the parser as critical is missing."},{status:700,message:"Calculation did not reach the end of execution."},{status:701,message:"The electronic structure is not converged."},{status:702,message:"The ionic relaxation is not converged."},{status:703,message:"VASP calculation encountered a critical error: {error_message}."},{status:704,message:"Outputs for diagnosis are missing, please make sure the `neb_data` and `run_status` quantities are requested for parsing."},{status:1001,message:"parsing an object has failed."},{status:1001,message:"parsing a file has failed."},{status:1002,message:"the parser is not able to parse the {quantity} quantity"},{status:1003,message:"the vasprun.xml was truncated and recovery parsing failed to parse at least one of the requested quantities: {quantities}, very likely the VASP calculation did not run properly"},{status:1004,message:"the parser is not able to compose one or more output nodes: {nodes}"},{status:1005,message:"Overflow detected in XML while parsing."}]},class:"aiida_vasp.calcs.neb:VaspNEBCalculation"},"vasp.vasp":{description:["General-purpose VASP calculation.",""," ---------------------------------"," By default retrieves only the 'OUTCAR', 'vasprun.xml', 'EIGENVAL', 'DOSCAR'"," and Wannier90 input / output objects. These objects are deleted after parsing."," Additional retrieve objects can be specified via the"," ``settings['ADDITIONAL_RETRIEVE_TEMPORARY_LIST']`` input. In addition, if you want to keep"," any objects after parsing, put them in ``settings['ADDITIONAL_RETRIEVE_LIST']`` which is empty"," by default.",""," Floating point precision for writing POSCAR objects can be adjusted using"," ``settings['poscar_precision']``, default: 10",""," The following assumes you are familiar with the AiiDA data structures and"," how to set up and run an AiiDA calculation in general.",""," Example usage::",""," from aiida.orm import CalculationFactory, DataFactory"," from aiida.work import submit",""," proc = CalculationFactory('vasp.vasp').process()"," inputs = proc.get_inputs_template()"," inputs.parameter = "," inputs.structure = "," inputs.kpoints = "," inputs.settings = "," inputs.potential = DataFactory('vasp.potcar').get_potcars_from_structure(structure, ...)"," inputs.code = ",""," submit(proc, **inputs)",""," Which is very similar to the workchain example.",""," Since we do not want the content parsers to know about the AiiDA infrastructure,"," i.e. processes etc. we have no access to the exit codes defined on the CalcJob."," We thus have to deal with failures in parsing directly in the write calls here."],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The kpoints to use (KPOINTS)."},{name:"parameters",required:!0,valid_types:"Dict",info:"The VASP input parameters (INCAR)."},{name:"potential",required:!0,valid_types:"PotcarData",info:"The potentials (POTCAR)."},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR)."},{name:"charge_density",required:!1,valid_types:"ChargedensityData, NoneType",info:"The charge density. (CHGCAR)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"dynamics",required:!1,valid_types:"Dict, NoneType",info:"The VASP parameters related to ionic dynamics, e.g. flags to set the selective dynamics"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:"A remote folder to restart from if need be"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters not related to VASP itself."},{name:"wavefunctions",required:!1,valid_types:"WavefunData, NoneType",info:"The wave function coefficients. (WAVECAR)"}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:333,message:"VASP did not produce any output and did likely not execute properly."},{status:350,message:"the retrieved folder data node could not be accessed."},{status:351,message:"the retrieved_temporary folder data node could not be accessed."},{status:352,message:"an object that is marked by the parser as critical is missing."},{status:700,message:"Calculation did not reach the end of execution."},{status:701,message:"The electronic structure is not converged."},{status:702,message:"The ionic relaxation is not converged."},{status:703,message:"VASP calculation encountered a critical error: {error_message}."},{status:704,message:"Outputs for diagnosis are missing, please make sure `run_status` and `notifications` quantities are requested for parsing."},{status:1001,message:"parsing an object has failed."},{status:1002,message:"the parser is not able to parse the {quantity} quantity"},{status:1003,message:"the vasprun.xml was truncated and recovery parsing failed to parse at least one of the requested quantities: {quantities}, very likely the VASP calculation did not run properly"},{status:1004,message:"the parser is not able to compose one or more output nodes: {nodes}"},{status:1005,message:"Overflow detected in XML while parsing."}]},class:"aiida_vasp.calcs.vasp:VaspCalculation"},"vasp.vasp2w90":{description:["General purpose Calculation for using vasp with the vasp2wannier90 interface."],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The kpoints to use (KPOINTS)."},{name:"parameters",required:!0,valid_types:"Dict",info:"The VASP input parameters (INCAR)."},{name:"potential",required:!0,valid_types:"PotcarData",info:"The potentials (POTCAR)."},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR)."},{name:"charge_density",required:!1,valid_types:"ChargedensityData, NoneType",info:"The charge density. (CHGCAR)"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"dynamics",required:!1,valid_types:"Dict, NoneType",info:"The VASP parameters related to ionic dynamics, e.g. flags to set the selective dynamics"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:"A remote folder to restart from if need be"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters not related to VASP itself."},{name:"wannier_parameters",required:!1,valid_types:"Dict, NoneType",info:"Input parameters for the Wannier90 interface."},{name:"wannier_projections",required:!1,valid_types:"OrbitalData, List, NoneType",info:"Projections to be defined in the Wannier90 input."},{name:"wavefunctions",required:!1,valid_types:"WavefunData, NoneType",info:"The wave function coefficients. (WAVECAR)"}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:333,message:"VASP did not produce any output and did likely not execute properly."},{status:350,message:"the retrieved folder data node could not be accessed."},{status:351,message:"the retrieved_temporary folder data node could not be accessed."},{status:352,message:"an object that is marked by the parser as critical is missing."},{status:700,message:"Calculation did not reach the end of execution."},{status:701,message:"The electronic structure is not converged."},{status:702,message:"The ionic relaxation is not converged."},{status:703,message:"VASP calculation encountered a critical error: {error_message}."},{status:704,message:"Outputs for diagnosis are missing, please make sure `run_status` and `notifications` quantities are requested for parsing."},{status:1001,message:"parsing an object has failed."},{status:1002,message:"the parser is not able to parse the {quantity} quantity"},{status:1003,message:"the vasprun.xml was truncated and recovery parsing failed to parse at least one of the requested quantities: {quantities}, very likely the VASP calculation did not run properly"},{status:1004,message:"the parser is not able to compose one or more output nodes: {nodes}"},{status:1005,message:"Overflow detected in XML while parsing."}]},class:"aiida_vasp.calcs.vasp2w90:Vasp2w90Calculation"}},"aiida.cmdline.data":{"vasp-potcar":"aiida_vasp.commands.potcar:potcar"},"aiida.data":{"vasp.archive":"aiida_vasp.data.archive:ArchiveData","vasp.chargedensity":"aiida_vasp.data.chargedensity:ChargedensityData","vasp.potcar":"aiida_vasp.data.potcar:PotcarData","vasp.potcar_file":"aiida_vasp.data.potcar:PotcarFileData","vasp.wavefun":"aiida_vasp.data.wavefun:WavefunData"},"aiida.groups":{"vasp.potcar":"aiida_vasp.data.potcar:PotcarGroup"},"aiida.parsers":{"vasp.neb":"aiida_vasp.parsers.neb:VtstNebParser","vasp.vasp":"aiida_vasp.parsers.vasp:VaspParser","vasp.vasp2w90":"aiida_vasp.parsers.vasp2w90:Vasp2w90Parser"},"aiida.workflows":{"vasp.bands":{description:["Extract the band structure using k-point paths fetched from SeeKpath."],spec:{inputs:[{name:"bands",required:!0,valid_types:"",info:""},{name:"code",required:!0,valid_types:"Code",info:""},{name:"dynamics",required:!0,valid_types:"",info:""},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"potential_family",required:!0,valid_types:"Str",info:""},{name:"potential_mapping",required:!0,valid_types:"Dict",info:""},{name:"restart_folder",required:!0,valid_types:"RemoteData",info:` + The folder to restart in, which contains the outputs from the prerun to extract the charge density. + `},{name:"smearing",required:!0,valid_types:"",info:""},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:""},{name:"chgcar",required:!1,valid_types:"ChargedensityData, NoneType",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:` + If True, clean the work dir upon the completion of a successful calculation. + `},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parameters",required:!1,valid_types:"Dict, NoneType",info:""},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"site_magnetization",required:!1,valid_types:"Dict, NoneType",info:"Site magnetization to be used as MAGMOM"},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `},{name:"wavecar",required:!1,valid_types:"WavefunData, NoneType",info:""}],outputs:[{name:"bands",required:!0,valid_types:"BandsData",info:""},{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"no called workchain detected"},{status:500,message:"unknown error detected in the bands workchain"},{status:2001,message:"BandsData not found in exposed_outputs"}]},class:"aiida_vasp.workchains.bands:BandsWorkChain"},"vasp.converge":{description:["A workchain to perform convergence tests."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"converge",required:!0,valid_types:"",info:""},{name:"dynamics",required:!0,valid_types:"",info:""},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"potential_family",required:!0,valid_types:"Str",info:""},{name:"potential_mapping",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:""},{name:"chgcar",required:!1,valid_types:"ChargedensityData, NoneType",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:` + If True, clean the work dir upon the completion of a successful calculation. + `},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:""},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax",required:!1,valid_types:"",info:""},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:` + The restart folder from a previous workchain run that is going to be used. + `},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"site_magnetization",required:!1,valid_types:"Dict, NoneType",info:"Site magnetization to be used as MAGMOM"},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `},{name:"wavecar",required:!1,valid_types:"WavefunData, NoneType",info:""}],outputs:[{name:"converge",required:!0,valid_types:"",info:""},{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"relax",required:!0,valid_types:"",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"no called workchain detected"},{status:500,message:"unknown error detected in the converge workchain"}]},class:"aiida_vasp.workchains.converge:ConvergeWorkChain"},"vasp.immigrant":{description:["Import a VASP run executed in the directory specified by folder_path."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:` + If True, clean the work dir upon the completion of a successful calculation. + `},{name:"folder_path",required:!1,valid_types:"Str, NoneType",info:"Deprecated."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"options",required:!1,valid_types:"Dict, NoneType",info:""},{name:"potential_family",required:!1,valid_types:"Str, NoneType",info:""},{name:"potential_mapping",required:!1,valid_types:"Dict, NoneType",info:""},{name:"remote_workdir",required:!1,valid_types:"str, NoneType",info:""},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"use_chgcar",required:!1,valid_types:"Bool, NoneType",info:` + If True, WavefunData (of WAVECAR) is attached. + `},{name:"use_wavecar",required:!1,valid_types:"Bool, NoneType",info:` + If True, WavefunData (of WAVECAR) is attached. + `},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_vasp.workchains.immigrant:VaspImmigrantWorkChain"},"vasp.master":{description:["The master workchain that selects sub workchains to perform necessary calculations."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"converge",required:!0,valid_types:"",info:""},{name:"dos",required:!0,valid_types:"",info:""},{name:"dynamics",required:!0,valid_types:"",info:""},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"potential_family",required:!0,valid_types:"Str",info:""},{name:"potential_mapping",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:""},{name:"chgcar",required:!1,valid_types:"ChargedensityData, NoneType",info:""},{name:"extract_bands",required:!1,valid_types:"Bool, NoneType",info:"Do you want to extract the band structure?"},{name:"extract_dos",required:!1,valid_types:"Bool, NoneType",info:"Do you want to extract the density of states?"},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:""},{name:"kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"The maximum distance between k-points in inverse AA."},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"relax",required:!1,valid_types:"",info:""},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:` + The restart folder from a previous workchain run that is going to be used. + `},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"site_magnetization",required:!1,valid_types:"Dict, NoneType",info:"Site magnetization to be used as MAGMOM"},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `},{name:"wavecar",required:!1,valid_types:"WavefunData, NoneType",info:""}],outputs:[{name:"bands",required:!1,valid_types:"",info:""},{name:"dos",required:!1,valid_types:"",info:""}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"no called workchain detected"},{status:500,message:"unknown error detected in the master workchain"}]},class:"aiida_vasp.workchains.master:MasterWorkChain"},"vasp.neb":{description:["The NEB workchain.",""," -------------------"," Error handling enriched wrapper around VaspNEBCalculation.",""," Deliberately conserves most of the interface (required inputs) of the VaspNEBCalculation class, but"," makes it possible for a user to interact with a workchain and not a calculation.",""," In addition, implement restarts of calculation when the calculation is net full converged for error handling."],spec:{inputs:[{name:"dynamics",required:!0,valid_types:"",info:""},{name:"final_structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR) for the final image."},{name:"initial_structure",required:!0,valid_types:"StructureData, CifData",info:"The input structure (POSCAR) for initial image."},{name:"neb_images",required:!0,valid_types:"StructureData, CifData",info:"Starting structure for the NEB images"},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:"The VASP input parameters (INCAR)."},{name:"potential_family",required:!0,valid_types:"Str",info:""},{name:"potential_mapping",required:!0,valid_types:"Dict",info:""},{name:"charge_density",required:!1,valid_types:"ChargedensityData",info:"The charge density. (CHGCAR)"},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:` + If True, clean the work dir upon the completion of a successful calculation. + `},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:""},{name:"kpoints_spacing",required:!1,valid_types:"Float, NoneType",info:"Spacing for the kpoints in units A^-1 * 2pi (CASTEP style `kpoints_mp_spacing`)"},{name:"kpoints_spacing_vasp",required:!1,valid_types:"Float, NoneType",info:"Spacing for the kpoints in units A^-1 (VASP style)"},{name:"ldau_mapping",required:!1,valid_types:"Dict, NoneType",info:"Mappings, see the doc string of 'get_ldau_keys'"},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:"A remote folder to restart from if need be"},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional parameters not related to VASP itself."},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `},{name:"wavefunctions",required:!1,valid_types:"WavefunData",info:"The wave function coefficients. (WAVECAR)"}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"Per-image misc output."},{name:"neb_misc",required:!0,valid_types:"Dict",info:"NEB related data combined for each image"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"structure",required:!0,valid_types:"StructureData",info:"NEB images"},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"Kpoints for each image."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization for each image."},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output file containing the plane wave coefficients."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."},{status:501,message:"Unrecoverable error in launched NEB calculations."},{status:700,message:"the user did not supply a potential family name"},{status:701,message:"ValueError was returned from get_potcars_from_structure"},{status:702,message:"the potential does not exist"},{status:703,message:"the exception: {exception} was thrown while massaging the parameters"}]},class:"aiida_vasp.workchains.neb:VaspNEBWorkChain"},"vasp.relax":{description:["Structure relaxation workchain."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dynamics",required:!0,valid_types:"",info:""},{name:"kpoints",required:!0,valid_types:"KpointsData",info:""},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"potential_family",required:!0,valid_types:"Str",info:""},{name:"potential_mapping",required:!0,valid_types:"Dict",info:""},{name:"relax",required:!0,valid_types:"",info:""},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:""},{name:"chgcar",required:!1,valid_types:"ChargedensityData, NoneType",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:` + If True, clean the work dir upon the completion of a successful calculation. + `},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:` + The restart folder from a previous workchain run that is going to be used. + `},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"site_magnetization",required:!1,valid_types:"Dict, NoneType",info:"Site magnetization to be used as MAGMOM"},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `},{name:"wavecar",required:!1,valid_types:"WavefunData, NoneType",info:""}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"relax",required:!0,valid_types:"",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"the called workchain does not contain the necessary relaxed output structure"},{status:420,message:"no called workchain detected"},{status:500,message:"unknown error detected in the relax workchain"},{status:502,message:"there was an error overriding the parameters"}]},class:"aiida_vasp.workchains.relax:RelaxWorkChain"},"vasp.vasp":{description:["The VASP workchain.",""," -------------------"," Error handling enriched wrapper around VaspCalculation.",""," Deliberately conserves most of the interface (required inputs) of the VaspCalculation class, but"," makes it possible for a user to interact with a workchain and not a calculation.",""," This is intended to be used instead of directly submitting a VaspCalculation,"," so that future features like"," automatic restarting, error checking etc. can be propagated to higher level workchains"," automatically by implementing them here.",""," Handlers are implemented to try fix common problems and improves the robustness."," Individual handlers can be enabled/disabled by setting the ``handler_overrides`` input port.",' Additional settings may be passed under the "settings" input, which is also forwarded to the'," calculations. The available options are:",""," - ``USE_WAVECAR_FOR_RESTART`` wether calculation restarts should use the WAVECAR. The default is ``True``.",""," Usage::",""," from aiida.common.extendeddicts import AttributeDict"," from aiida.work import submit"," basevasp = WorkflowFactory('vasp.vasp')"," inputs = basevasp.get_builder()"," inputs = AttributeDict()"," ## ... set inputs"," submit(basevasp, **inputs)",""," To see a working example, including generation of input nodes from scratch, please"," refer to ``examples/run_vasp_lean.py``."],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:""},{name:"dynamics",required:!0,valid_types:"",info:""},{name:"kpoints",required:!0,valid_types:"KpointsData",info:""},{name:"options",required:!0,valid_types:"Dict",info:""},{name:"parameters",required:!0,valid_types:"Dict",info:""},{name:"potential_family",required:!0,valid_types:"Str",info:""},{name:"potential_mapping",required:!0,valid_types:"Dict",info:""},{name:"structure",required:!0,valid_types:"StructureData, CifData",info:""},{name:"chgcar",required:!1,valid_types:"ChargedensityData, NoneType",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool, NoneType",info:` + If True, clean the work dir upon the completion of a successful calculation. + `},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int, NoneType",info:` + The maximum number of iterations to perform. + `},{name:"metadata",required:!1,valid_types:"",info:""},{name:"restart_folder",required:!1,valid_types:"RemoteData, NoneType",info:` + The restart folder from a previous workchain run that is going to be used. + `},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:""},{name:"site_magnetization",required:!1,valid_types:"Dict, NoneType",info:"Site magnetization to be used as MAGMOM"},{name:"verbose",required:!1,valid_types:"Bool, NoneType",info:` + If True, enable more detailed output during workchain execution. + `},{name:"wavecar",required:!1,valid_types:"WavefunData, NoneType",info:""}],outputs:[{name:"misc",required:!0,valid_types:"Dict",info:"The output parameters containing smaller quantities that do not depend on system size."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:"The output band structure."},{name:"born_charges",required:!1,valid_types:"ArrayData",info:"The output Born effective charges."},{name:"charge_density",required:!1,valid_types:"ArrayData",info:"The output charge density."},{name:"chgcar",required:!1,valid_types:"ChargedensityData",info:"The output charge density CHGCAR file."},{name:"custom_outputs",required:!1,valid_types:"",info:""},{name:"dielectrics",required:!1,valid_types:"ArrayData",info:"The output dielectric functions."},{name:"dos",required:!1,valid_types:"ArrayData",info:"The output dos."},{name:"dynmat",required:!1,valid_types:"ArrayData",info:"The output dynamical matrix."},{name:"energies",required:!1,valid_types:"ArrayData",info:"The output total energies."},{name:"forces",required:!1,valid_types:"ArrayData",info:"The output forces."},{name:"hessian",required:!1,valid_types:"ArrayData",info:"The output Hessian matrix."},{name:"kpoints",required:!1,valid_types:"KpointsData",info:"The output k-points."},{name:"magnetization_density",required:!1,valid_types:"ArrayData",info:"The output magnetization density."},{name:"projectors",required:!1,valid_types:"ArrayData",info:"The output projectors of decomposition."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."},{name:"site_magnetization",required:!1,valid_types:"Dict",info:"The output of the site magnetization"},{name:"stress",required:!1,valid_types:"ArrayData",info:"The output stress."},{name:"structure",required:!1,valid_types:"StructureData",info:"The output structure."},{name:"trajectory",required:!1,valid_types:"TrajectoryData",info:"The output trajectory data."},{name:"wavecar",required:!1,valid_types:"WavefunData",info:"The output plane wave coefficients file."}],exit_codes:[{status:0,message:"the sun is shining"},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:300,message:"the calculation is missing at least one required output in the restart workchain"},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:400,message:"the run_calculation step did not successfully add a calculation node to the context"},{status:401,message:"the maximum number of iterations was exceeded"},{status:402,message:"the calculation finished with an unexpected calculation state"},{status:403,message:"the calculation experienced and unexpected failure"},{status:404,message:"the calculation failed to submit, twice in a row"},{status:405,message:"the calculation failed for an unknown reason, twice in a row"},{status:500,message:"Missing critical output for inspecting the status of the calculation."},{status:501,message:"Cannot handle the error - inputs are likely need to be revised manually. Message: {message}"},{status:502,message:"Cannot handle the error - the last calculation did not reach the end of execution."},{status:503,message:"Cannot handle the error - the last calculation did not reach electronic convergence."},{status:504,message:"The ionic relaxation is not converged."},{status:505,message:"At least one of the ionic steps during the relaxation has did not have converged electronic structure."},{status:700,message:"the user did not supply a potential family name"},{status:701,message:"ValueError was returned from get_potcars_from_structure"},{status:702,message:"the potential does not exist"},{status:703,message:"the exception: {exception} was thrown while massaging the parameters"}]},class:"aiida_vasp.workchains.vasp:VaspWorkChain"}},console_scripts:{"mock-vasp":"aiida_vasp.commands.mock_vasp:mock_vasp","mock-vasp-strict":"aiida_vasp.commands.mock_vasp:mock_vasp_strict"}},commits_count:199,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:4},{colorclass:"brown",text:"Parsers",count:3},{colorclass:"red",text:"Data",count:5},{colorclass:"green",text:"Workflows",count:7},{colorclass:"purple",text:"Console scripts",count:2},{colorclass:"orange",text:"Other (Data commands, Groups)",count:2}],pip_install_cmd:"pip install aiida-vasp",is_installable:"True"},"aiida-vibroscopy":{entry_point_prefix:"vibroscopy",plugin_info:"https://raw.githubusercontent.com/bastonero/aiida-vibroscopy/main/pyproject.toml",code_home:"https://github.com/bastonero/aiida-vibroscopy",version_file:"https://raw.githubusercontent.com/bastonero/aiida-vibroscopy/main/src/aiida_vibroscopy/__init__.py",pip_url:"aiida-vibroscopy",documentation_url:"https://aiida-vibroscopy.readthedocs.io/en/latest/",name:"aiida-vibroscopy",package_name:"aiida_vibroscopy",hosted_on:"github.com",metadata:{release_date:"2024-09-11",description:"AiiDA plugin for vibrational spectoscopy using Quantum ESPRESSO.",author_email:"Lorenzo Bastonero ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: Other/Proprietary License","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"1.1.1"},aiida_version:">=2.2.2,<3.0.0",entry_points:{"aiida.data":{"vibroscopy.fp":"aiida_vibroscopy.data.vibro_fp:VibrationalFrozenPhononData","vibroscopy.vibrational":"aiida_vibroscopy.data.vibro_lr:VibrationalData"},"aiida.workflows":{"vibroscopy.dielectric":{description:["Workchain computing different second and third order tensors.",""," It computes the high frequency dielectric tensor, the Born effective charges,"," the non-linear optical susceptibility and Raman tensors"," using homogeneous small electric fields via the electric enthalpy functional."],spec:{inputs:[{name:"central_difference",required:!0,valid_types:"",info:"The inputs for the central difference scheme."},{name:"property",required:!0,valid_types:"str",info:`Valid inputs are:ir + * born-charges + * dielectric + * nac + * bec + * raman + * susceptibility-derivative + * non-linear-susceptibility`},{name:"scf",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` that will be used to run the electric enthalpy scfs."},{name:"settings",required:!0,valid_types:"",info:"Options for how to run the workflow."},{name:"symmetry",required:!0,valid_types:"",info:"Namespace for symmetry related inputs."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"kpoints_parallel_distance",required:!1,valid_types:"Float, NoneType",info:"Distance of the k-points in reciprocal space along the parallel direction of each applied electric field."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parent_scf",required:!1,valid_types:"RemoteData, NoneType",info:"Scf parent folder from where restarting the scfs with electric fields."}],outputs:[{name:"fields_data",required:!0,valid_types:"",info:"Namespace for passing TrajectoryData containing forces and polarization."},{name:"tensors",required:!0,valid_types:"ArrayData",info:"Contains high frequency dielectric and Born effectivecharges tensors computed in Cartesian coordinates. Depending on the inputs, it can also contain the derivatives of the susceptibility in respect to the atomic positions (called `Raman tensors`) and the non linear optical susceptibility, always expressed in Cartesian coordinates."},{name:"accuracy_order",required:!1,valid_types:"Int",info:""},{name:"critical_electric_field",required:!1,valid_types:"Float",info:""},{name:"electric_field_step",required:!1,valid_types:"Float",info:""},{name:"units",required:!1,valid_types:"Dict",info:"Units of the susceptibility derivatives tensors."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:400,message:"The initial scf work chain failed."},{status:401,message:"The nscf work chain failed."},{status:402,message:"The electric field scf work chain failed for direction {direction}."},{status:403,message:"The numerical derivatives calculation failed."},{status:404,message:"The scf PwBaseWorkChain sub process in iteration returned a non integer total magnetization (threshold exceeded)."}]},class:"aiida_vibroscopy.workflows.dielectric.base:DielectricWorkChain"},"vibroscopy.dielectric.numerical_derivatives":{description:["Workchain carrying out numerical derivatives.",""," It computes the first and second order derivatives"," of forces and polarization in respect to electric field,"," to obtain dielectric tensor, Born effective charges,"," non linear optical susceptibility and Raman tensors.",""," Forces and polarization must be passed as TrajectoryData"," as a dictionary in `data`. Numerical derivatives can have"," different number of evaluation points, depending on order and accuracy."," The price to pay is the standardization of the structure of"," the dictionary to pass to this namespace.",""," To understand, let's review the approach.In central differencs approach"," we need the evaluation of the function at the value we want"," the derivative (in our case at :math:`\\\\mathcal{E}=0`,"," E is the electric field), and at"," displaced positions from this value."," The evaluation of the function at these points will"," have weights (or coefficients), which depend on order and accuracy."," For example:",""," - :math:`\\\\frac{df}{dx} = \\\\frac{ 0.5 \\\\cdot f(+1.0 \\\\cdot h) -0.5 \\\\cdot f(-1.0 \\\\cdot h) }{h} +\\\\mathcal{O}(h^2)`"," - :math:`\\\\frac{d^2 f}{dx^2} = \\\\frac{ 1.0 \\\\cdot f(+1.0 \\\\cdot h) -2.0 \\\\cdot f(0. \\\\cdot h) +1.0 \\\\cdot f(-1.0 \\\\cdot h) }{h^2} +\\\\mathcal{O}(h^2)`",""," Referring to the coefficients for each step as :math:`c_i`,"," where `i` is an integer, our convention is"," to put in sequence the Trajectory data with increasing"," numbers as labels, for example:",""," | '0': TrajectoryData for :math:`c_1`,"," | '1': TrajectoryData for :math:`c_{-1}`,"," | '2': TrajectoryData for :math:`c_2`,"," | '3': TrajectoryData for :math:`c_{-2}`,"," | ...",""," This way to creating an analogous of an array with"," coefficients :math:`[c_1,c_{-1},c_2,c_{-2}, \\\\dots]`.",""," These dictionaries are going to be put as sub-dictionary"," in a general `data` dictionary. Each sub-dict"," has to be put with a key with suffix a number indicating"," which tensor component is referring to."," In our case, we use a similar Voigt notation."," Namely we have two cases:",""," * first order derivatives: keys suffices are 0,1,2;"," 0 for :math:`[i,x]`, 1 for :math:`[i,y]`, 2 for"," :math:`[i,z]` (with :math:`i={x,y,z}`)"," * second order derivatives: keys suffices are 0,...5;"," 0 for :math:`[i,x,x]`, :math:`\\dots` (as in Voigt),"," 5 for :math:`[i,x,y]` (with :math:`i={x,y,z}`)",""," The prefix can be anything. Best practice is using ``field_``"," with and underscorre as prefix. The Trajectory data for the"," :math:`c_0` coefficient (i.e. the one with :math:`\\mathcal{E}=0`)"," must be passed with a different key, namely ``null_field``."," This is to avoid errors and due to the fact that is common"," to the all derivatives."],spec:{inputs:[{name:"central_difference",required:!0,valid_types:"",info:"The inputs for the central difference scheme."},{name:"data",required:!0,valid_types:"",info:"Namespace for passing TrajectoryData containing forces and polarization."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"symmetry",required:!0,valid_types:"",info:""},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"tensors",required:!0,valid_types:"ArrayData",info:"Contains high frequency dielectric and Born effectivecharges tensors computed in Cartesian coordinates. Depending on the inputs, it can also contain the derivatives of the susceptibility in respect to the atomic positions (called `Raman tensors`) and the non linear optical susceptibility, always expressed in Cartesian coordinates."},{name:"units",required:!1,valid_types:"Dict",info:"Units of the susceptibility derivatives tensors."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_vibroscopy.workflows.dielectric.numerical_derivatives:NumericalDerivativesWorkChain"},"vibroscopy.phonons.harmonic":{description:["Workchain for frozen phonons calculations.",""," Non-analytical constants (NAC) and higher order mixed derivatives are computed"," via finite differences through finite electric fields."," See :class:`~aiida_vibroscopy.workflows.DielectricWorkChain`"," for more details on how they are carried out."],spec:{inputs:[{name:"phonon",required:!0,valid_types:"Data",info:"Inputs for the `PhononWorkChain` that will beused to calculate the force constants."},{name:"settings",required:!0,valid_types:"",info:"Options for how to run the workflow."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"symmetry",required:!0,valid_types:"",info:"Namespace for symmetry related inputs."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"dielectric",required:!1,valid_types:"Data",info:"Inputs for the `DielectricWorkChain` that will beused to calculate the mixed derivatives with electric field."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"phonopy",required:!1,valid_types:"Data",info:"Inputs for the `PhonopyCalculation` that willbe used to calculate the inter-atomic force constants, or for post-processing."}],outputs:[{name:"output_phonon",required:!0,valid_types:"",info:"Outputs of the `PhononWorkChain`."},{name:"vibrational_data",required:!0,valid_types:"VibrationalData, VibrationalFrozenPhononData",info:"The phonopy data with supercells displacements, forces and (optionally)nac parameters to use in the post-processing calculation."},{name:"output_dielectric",required:!1,valid_types:"",info:"Outputs of the `DielectricWorkChain`."},{name:"output_phonopy",required:!1,valid_types:"",info:"Outputs of the post-processing via `phonopy`."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:400,message:"The phonon workchain failed."},{status:401,message:"The dielectric workchain failed."},{status:402,message:"The phonopy calculation failed."}]},class:"aiida_vibroscopy.workflows.phonons.harmonic:HarmonicWorkChain"},"vibroscopy.phonons.phonon":{description:["Class for computing force constants of phonons, without non-analytical corrections."],spec:{inputs:[{name:"scf",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` that will be used to run the electric enthalpy scfs."},{name:"settings",required:!0,valid_types:"",info:"Options for how to run the workflow."},{name:"symmetry",required:!0,valid_types:"",info:"Namespace for symmetry related inputs."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"displacement_generator",required:!1,valid_types:"Dict, NoneType",info:`Info for displacements generation. The following flags are allowed: + distance + is_plusminus + is_diagonal + is_trigonal + number_of_snapshots + random_seed + cutoff_frequency`},{name:"metadata",required:!1,valid_types:"",info:""},{name:"phonopy",required:!1,valid_types:"Data",info:"Inputs for the `PhonopyCalculation` that willbe used to calculate the inter-atomic force constants, or for post-processing."},{name:"primitive_matrix",required:!1,valid_types:"List, NoneType",info:"Primitive matrix that defines the primitive cell from the unitcell."},{name:"supercell_matrix",required:!1,valid_types:"List, NoneType",info:"Supercell matrix that defines the supercell from the unitcell."}],outputs:[{name:"phonopy_data",required:!0,valid_types:"PhonopyData",info:"The phonopy data with supercells displacements, forces to use in the post-processing calculation."},{name:"supercells_forces",required:!0,valid_types:"ArrayData, TrajectoryData",info:"The forces acting on the atoms of each supercell."},{name:"output_phonopy",required:!1,valid_types:"",info:""},{name:"supercells",required:!1,valid_types:"StructureData",info:"The supercells with displacements."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:400,message:"The initial supercell scf work chain failed."},{status:401,message:"The initial PwBaseWorkChain sub process returned a non integer total magnetization."},{status:402,message:"At least one sub processe did not finish successfully."},{status:403,message:"The phonopy calculation did not finish correctly."}]},class:"aiida_vibroscopy.workflows.phonons.base:PhononWorkChain"},"vibroscopy.spectra.intensities_average":{description:["Workchain that computes IR and Raman spatial and q-direction average spectra."],spec:{inputs:[{name:"vibrational_data",required:!0,valid_types:"VibrationalData, VibrationalFrozenPhononData",info:"Vibrational data containing force constants or frozen phonons forces, nac parameters and/or susceptibility derivatives."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"parameters",required:!1,valid_types:"Dict",info:"Options for averaging on the non-analytical directions."}],outputs:[{name:"ir_averaged",required:!0,valid_types:"ArrayData",info:"Contains high frequency dielectric tensor computed in Cartesian coordinates."},{name:"raman_averaged",required:!1,valid_types:"ArrayData",info:"Contains Born effective charges tensors computed in Cartesian coordinates."},{name:"units",required:!1,valid_types:"Dict",info:"Units of intensities and frequencies."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_vibroscopy.workflows.spectra.intensities_average:IntensitiesAverageWorkChain"},"vibroscopy.spectra.iraman":{description:["Workchain for automatically compute IR and Raman spectra using finite displacements and fields.",""," For other details of the sub-workchains used, see also:"," * :class:`~aiida_vibroscopy.workflows.dielectric.base.DielectricWorkChain` for finite fields"," * :class:`~aiida_vibroscopy.workflows.phonons.base.PhononWorkChain` for finite displacements"],spec:{inputs:[{name:"dielectric",required:!0,valid_types:"Data",info:"Inputs for the `DielectricWorkChain` that will beused to calculate the mixed derivatives with electric field."},{name:"phonon",required:!0,valid_types:"Data",info:"Inputs for the `PhononWorkChain` that will beused to calculate the force constants."},{name:"settings",required:!0,valid_types:"",info:"Options for how to run the workflow."},{name:"structure",required:!0,valid_types:"StructureData",info:""},{name:"symmetry",required:!0,valid_types:"",info:"Namespace for symmetry related inputs."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"intensities_average",required:!1,valid_types:"Data",info:"Inputs for the `IntensitiesAverageWorkChain` that willbe used to run the average calculation over intensities."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"output_phonon",required:!0,valid_types:"",info:"Outputs of the `PhononWorkChain`."},{name:"vibrational_data",required:!0,valid_types:"VibrationalData, VibrationalFrozenPhononData",info:"The phonopy data with supercells displacements, forces and (optionally)nac parameters to use in the post-processing calculation."},{name:"fake",required:!1,valid_types:"",info:""},{name:"output_dielectric",required:!1,valid_types:"",info:"Outputs of the `DielectricWorkChain`."},{name:"output_intensities_average",required:!1,valid_types:"",info:"Intensities average over space and q-points."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:400,message:"The averaging procedure for intensities had an unexpected error."},{status:401,message:"The averaging procedure for intensities had an unexpected error."}]},class:"aiida_vibroscopy.workflows.spectra.iraman:IRamanSpectraWorkChain"}}},commits_count:19,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"red",text:"Data",count:2},{colorclass:"green",text:"Workflows",count:6}],pip_install_cmd:"pip install aiida-vibroscopy",is_installable:"True"},"aiida-wannier90":{code_home:"https://github.com/aiidateam/aiida-wannier90",documentation_url:"https://aiida-wannier90.readthedocs.io/",entry_point_prefix:"wannier90",pip_url:"aiida-wannier90",plugin_info:"https://raw.github.com/aiidateam/aiida-wannier90/master/setup.json",name:"aiida-wannier90",package_name:"aiida_wannier90",hosted_on:"github.com",metadata:{release_date:"2024-05-10",description:"AiiDA Plugin for the Wannier90 code",author:"Junfeng Qiao, Dominik Gresch, Antimo Marrazzo, Daniel Marchand, Giovanni Pizzi, Norma Rivano, The AiiDA team",classifiers:["Development Status :: 5 - Production/Stable","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Physics"],version:"2.2.0"},aiida_version:">=2.0,<3",entry_points:{"aiida.calculations":{"wannier90.postw90":{description:["Plugin for Wannier90.",""," Wannier90 is a code for computing maximally-localized Wannier functions."," See http://www.wannier.org/ for more details."],spec:{inputs:[{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters for the Wannier90 code"},{name:"parent_folder",required:!0,valid_types:"RemoteData",info:"Get input files (``.amn``, ``.mmn``, ...) from a class ``RemoteData`` possibly stored in a remote computer."},{name:"structure",required:!0,valid_types:"StructureData",info:"input crystal structure"},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"A list of k-points along a path to be used for bands interpolation; it should contain `labels`. Specify either this or `kpoint_path`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"kpoint_path",required:!1,valid_types:"Dict, NoneType",info:"Description of the k-points path to be used for bands interpolation; it should contain two properties: a list ``path`` of length-2 tuples with the labels of the endpoints of the path; and a dictionary ``point_coords`` giving the scaled coordinates for each high-symmetry endpoint."},{name:"kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"k-point mesh used in the NSCF calculation."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"projections",required:!1,valid_types:"OrbitalData, Dict, List, NoneType",info:"Starting projections for the Wannierisation procedure."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional settings to manage the Wannier90 calculation."}],outputs:[{name:"boltzwann",required:!0,valid_types:"",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:"The ``output_parameters`` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"interpolated_bands",required:!1,valid_types:"BandsData",info:"The interpolated band structure by Wannier90 (if any)."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"The retrieved folder data node could not be accessed."},{status:210,message:"The retrieved folder did not contain the required stdout output file."},{status:300,message:"A Wannier90 error file (.werr) has been found."},{status:400,message:'The string "Exiting..." has been found in the Wannier90 output (some partial output might have been parsed).'},{status:401,message:"An error related to bvectors has been found in the Wannier90 output."},{status:402,message:"Energy window contains fewer states than number of target WFs."},{status:403,message:"Error plotting Wanier functions in cube format."},{status:404,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:405,message:"Some output files were missing probably because the calculation got interrupted."},{status:406,message:"The retrieved temporary folder could not be accessed."}]},class:"aiida_wannier90.calculations:Postw90Calculation"},"wannier90.wannier90":{description:["Plugin for Wannier90.",""," Wannier90 is a code for computing maximally-localized Wannier functions."," See http://www.wannier.org/ for more details."],spec:{inputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"k-point mesh used in the NSCF calculation."},{name:"parameters",required:!0,valid_types:"Dict",info:"Input parameters for the Wannier90 code"},{name:"structure",required:!0,valid_types:"StructureData",info:"input crystal structure"},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"A list of k-points along a path to be used for bands interpolation; it should contain `labels`. Specify either this or `kpoint_path`."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"kpoint_path",required:!1,valid_types:"Dict, NoneType",info:"Description of the k-points path to be used for bands interpolation; it should contain two properties: a list ``path`` of length-2 tuples with the labels of the endpoints of the path; and a dictionary ``point_coords`` giving the scaled coordinates for each high-symmetry endpoint."},{name:"local_input_folder",required:!1,valid_types:"FolderData, NoneType",info:"Get input files (``.amn``, ``.mmn``, ...) from a class ``FolderData`` stored in the AiiDA repository."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"projections",required:!1,valid_types:"OrbitalData, Dict, List, NoneType",info:"Starting projections for the Wannierisation procedure."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"remote_input_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Get input files (``.amn``, ``.mmn``, ...) from a class ``RemoteData`` possibly stored in a remote computer."},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional settings to manage the Wannier90 calculation."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The ``output_parameters`` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"interpolated_bands",required:!1,valid_types:"BandsData",info:"The interpolated band structure by Wannier90 (if any)."},{name:"nnkp_file",required:!1,valid_types:"SinglefileData",info:"The ``.nnkp`` file, produced only in -pp (postproc) mode."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:200,message:"The retrieved folder data node could not be accessed."},{status:210,message:"The retrieved folder did not contain the required stdout output file."},{status:300,message:"A Wannier90 error file (.werr) has been found."},{status:400,message:'The string "Exiting..." has been found in the Wannier90 output (some partial output might have been parsed).'},{status:401,message:"An error related to bvectors has been found in the Wannier90 output."},{status:402,message:"Energy window contains fewer states than number of target WFs."},{status:403,message:"Error plotting Wanier functions in cube format."},{status:404,message:"The stdout output file was incomplete probably because the calculation got interrupted."}]},class:"aiida_wannier90.calculations:Wannier90Calculation"}},"aiida.parsers":{"wannier90.postw90":"aiida_wannier90.parsers:Postw90Parser","wannier90.wannier90":"aiida_wannier90.parsers:Wannier90Parser"},"aiida.workflows":{"wannier90.minimal":{description:["Workchain to run a full stack of Quantum ESPRESSO + Wannier90 for GaAs.",""," Note that this is mostly to be used as an example, as there is no"," error checking and runs directly Quantum ESPRESSO calculations rather"," than the base workflows."],spec:{inputs:[{name:"kpoint_path",required:!0,valid_types:"Dict",info:"The kpoints path for the NSCF run and Wannierisation."},{name:"kpoints_nscf",required:!0,valid_types:"KpointsData",info:"The kpoints for the NSCF run and Wannierisation."},{name:"kpoints_scf",required:!0,valid_types:"KpointsData",info:"The kpoints for the SCF run."},{name:"projections",required:!0,valid_types:"OrbitalData",info:"The projections for the Wannierisation."},{name:"pseudo_family",required:!0,valid_types:"Str",info:"The name of a pseudopotential family to use."},{name:"pw2wannier90_code",required:!0,valid_types:"Code",info:"The `pw2wannier90.x` code to use for the `Pw2Wannier90Calculation`s."},{name:"pw_code",required:!0,valid_types:"Code",info:"The `pw.x` code to use for the `PwCalculation`s."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"wannier_code",required:!0,valid_types:"Code",info:"The `wannier90.x` code to use for the `Wannier90Calculation`s."},{name:"max_wallclock_seconds",required:!1,valid_types:"Int, NoneType",info:"Maximum wallclock time in seconds"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"num_machines",required:!1,valid_types:"Int, NoneType",info:"The number of machines (nodes) to use"}],outputs:[{name:"matrices_folder",required:!0,valid_types:"FolderData",info:""},{name:"nnkp_file",required:!0,valid_types:"SinglefileData",info:""},{name:"nscf_output",required:!0,valid_types:"Dict",info:""},{name:"p2wannier_output",required:!0,valid_types:"Dict",info:""},{name:"pw2wan_remote_folder",required:!0,valid_types:"RemoteData",info:""},{name:"scf_output",required:!0,valid_types:"Dict",info:""},{name:"wannier_bands",required:!0,valid_types:"BandsData",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."}]},class:"aiida_wannier90.workflows.minimal:MinimalW90WorkChain"}}},commits_count:5,development_status:"stable",errors:[],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:2},{colorclass:"brown",text:"Parsers",count:2},{colorclass:"green",text:"Workflows",count:1}],pip_install_cmd:"pip install aiida-wannier90",is_installable:"True"},"aiida-wannier90-workflows":{code_home:"https://github.com/aiidateam/aiida-wannier90-workflows",development_status:"stable",entry_point_prefix:"wannier90_workflows",pip_url:"aiida-wannier90-workflows",plugin_info:"https://raw.github.com/aiidateam/aiida-wannier90-workflows/master/setup.json",name:"aiida-wannier90-workflows",package_name:"aiida_wannier90_workflows",hosted_on:"github.com",metadata:{release_date:"2024-10-01",description:"Advanced AiiDA workflows for Wannier90",author_email:"The AiiDA team ",classifiers:["Development Status :: 5 - Production/Stable","Environment :: Plugins","Framework :: AiiDA","Intended Audience :: Science/Research","License :: Free To Use But Restricted","Natural Language :: English","Operating System :: MacOS :: MacOS X","Operating System :: POSIX :: Linux","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.12","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Physics"],version:"2.3.0"},aiida_version:">=2.5,<3.0",entry_points:{"aiida.calculations":{"wannier90_workflows.split":{description:["AiiDA calculation plugin wrapping the split AMN/MMN/EIG script."],spec:{inputs:[{name:"num_val",required:!0,valid_types:"Int",info:"Number of valence WFs."},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"parent_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote folder containing win/amn/mmn/eig/unk files."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."},{name:"rotate_unk",required:!1,valid_types:"Bool",info:"Number of valence WFs."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"Output parameters."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"remote_folder_cond",required:!0,valid_types:"RemoteData",info:"Remote folder for conduction."},{name:"remote_folder_val",required:!0,valid_types:"RemoteData",info:"Remote folder for valence."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"Calculation did not produce all expected output files."},{status:301,message:"The retrieved temporary folder could not be accessed."}]},class:"aiida_wannier90_workflows.calculations.split:Wannier90SplitCalculation"}},"aiida.parsers":{"wannier90_workflows.split":"aiida_wannier90_workflows.parsers.split:Wannier90SplitParser"},"aiida.workflows":{"wannier90_workflows.bands":{description:["WorkChain to automatically compute a Wannier band structure for a given structure."],spec:{inputs:[{name:"pw2wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Pw2wannier90BaseWorkChain`."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Wannier90BaseWorkChain`."},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Explicit kpoints to use for the wannier90 bands interpolation. If specified, wannier90 will use this list of kpoints and will not use the `bands_num_points` mechanism to auto generate a list of kpoints along the kpath. If not specified, the workchain will run seekpath to generate a primitive cell and a bands_kpoints. Specify either this or `bands_kpoints` or `bands_kpoints_distance`. This ensures the wannier interpolated bands has the exact same number of kpoints as PW bands, to calculate bands distance."},{name:"bands_kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"Minimum kpoints distance for seekpath to generate a list of kpoints along the path. Specify either this or `bands_kpoints` or `kpoint_path`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If True, work directories of all called calculation will be cleaned at the end of execution."},{name:"kpoint_path",required:!1,valid_types:"KpointsData, NoneType",info:"High symmetry kpoints to use for the wannier90 bands interpolation. If specified, the high symmetry kpoint labels will be used and wannier90 will use the `bands_num_points` mechanism to auto generate a list of kpoints along the kpath. If not specified, the workchain will run seekpath to generate a primitive cell and a bands_kpoints. Specify either this or `bands_kpoints` or `bands_kpoints_distance`."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the NSCF calculation."},{name:"open_grid",required:!1,valid_types:"Data",info:"Inputs for the `OpenGridBaseWorkChain`, if not specified the open_grid step is skipped."},{name:"projwfc",required:!1,valid_types:"Data",info:"Inputs for the `ProjwfcBaseWorkChain`."},{name:"scf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the SCF calculation."}],outputs:[{name:"band_structure",required:!0,valid_types:"BandsData",info:"The Wannier interpolated band structure."},{name:"pw2wannier90",required:!0,valid_types:"",info:""},{name:"wannier90",required:!0,valid_types:"",info:""},{name:"wannier90_pp",required:!0,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"",info:""},{name:"open_grid",required:!1,valid_types:"",info:""},{name:"primitive_structure",required:!1,valid_types:"StructureData",info:"The normalized and primitivized structure for which the calculations are computed."},{name:"projwfc",required:!1,valid_types:"",info:""},{name:"scf",required:!1,valid_types:"",info:""},{name:"seekpath_parameters",required:!1,valid_types:"Dict",info:"The parameters used in the SeeKpath call to normalize the input or relaxed structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"the scf PwBaseWorkChain sub process failed"},{status:430,message:"the nscf PwBaseWorkChain sub process failed"},{status:440,message:"the ProjwfcBaseWorkChain sub process failed"},{status:450,message:"the postproc Wannier90BaseWorkChain sub process failed"},{status:460,message:"the Pw2wannier90BaseWorkChain sub process failed"},{status:470,message:"the Wannier90BaseWorkChain sub process failed"},{status:480,message:"outputs sanity check failed"},{status:490,message:"the OpenGridBaseWorkChain sub process failed"}]},class:"aiida_wannier90_workflows.workflows.bands:Wannier90BandsWorkChain"},"wannier90_workflows.base.open_grid":{description:["Workchain to run a open_grid calculation with automated error handling and restarts."],spec:{inputs:[{name:"open_grid",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"kpoints",required:!0,valid_types:"KpointsData",info:"The explicit list of kpoints of the unfolded kmesh"},{name:"kpoints_mesh",required:!0,valid_types:"KpointsData",info:"The dimensions of the unfolded kmesh"},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:311,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_wannier90_workflows.workflows.base.open_grid:OpenGridBaseWorkChain"},"wannier90_workflows.base.projwfc":{description:["Workchain to run a projwfc.x calculation with automated error handling and restarts."],spec:{inputs:[{name:"projwfc",required:!0,valid_types:"Data",info:""},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""}],outputs:[{name:"Dos",required:!0,valid_types:"XyData",info:""},{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"bands",required:!1,valid_types:"BandsData",info:""},{name:"bands_down",required:!1,valid_types:"BandsData",info:""},{name:"bands_up",required:!1,valid_types:"BandsData",info:""},{name:"projections",required:!1,valid_types:"ProjectionData",info:""},{name:"projections_down",required:!1,valid_types:"ProjectionData",info:""},{name:"projections_up",required:!1,valid_types:"ProjectionData",info:""},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:311,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_wannier90_workflows.workflows.base.projwfc:ProjwfcBaseWorkChain"},"wannier90_workflows.base.pw2wannier90":{description:["Workchain to run a pw2wannier90 calculation with automated error handling and restarts."],spec:{inputs:[{name:"pw2wannier90",required:!0,valid_types:"Data",info:""},{name:"bands",required:!1,valid_types:"BandsData, NoneType",info:"Bands to calculate SCDM `mu`, `sigma`."},{name:"bands_projections",required:!1,valid_types:"ProjectionData, NoneType",info:"Bands projectability to calculate SCDM `mu`, `sigma`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"scdm_sigma_factor",required:!1,valid_types:"Float",info:"The `sigma` factor of occupation function for SCDM projection."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:""},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:311,message:"The stdout output file was incomplete probably because the calculation got interrupted."},{status:400,message:"Error when fitting `scdm_mu` and `scdm_sigma`."},{status:401,message:"The maximum number of iterations was exceeded."},{status:402,message:"The process failed for an unknown reason, twice in a row."}]},class:"aiida_wannier90_workflows.workflows.base.pw2wannier90:Pw2wannier90BaseWorkChain"},"wannier90_workflows.base.wannier90":{description:["Workchain to run a `Wannier90Calculation` with automated error handling and restarts."],spec:{inputs:[{name:"wannier90",required:!0,valid_types:"Data",info:""},{name:"auto_energy_windows",required:!1,valid_types:"Bool",info:"If True use the energy corresponding to projectability = auto_energy_windows_threshold as `dis_froz_max` for wannier90."},{name:"auto_energy_windows_threshold",required:!1,valid_types:"Float",info:"Threshold for auto_energy_windows."},{name:"bands",required:!1,valid_types:"BandsData, NoneType",info:"For shift_energy_windows, if provided the energy windows will be shifted by Fermi energy for metals or minimum of lowest-unoccupied bands for insulators. The bands should be along a kpath to better estimate the band gap. For auto_energy_windows, the bands is used to find out the energy corresponds to projectability = auto_energy_windows_threshold, the energy is used as `dis_froz_max`. In this case the bands should be on a nscf kmesh."},{name:"bands_projections",required:!1,valid_types:"ProjectionData, NoneType",info:"Projectability of bands to auto set `dis_froz_max`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation jobs will be cleaned at the end of execution."},{name:"guiding_centres_projections",required:!1,valid_types:"OrbitalData, Dict, List, NoneType",info:"Projections block for `guiding_centres = True`."},{name:"handler_overrides",required:!1,valid_types:"Dict, NoneType",info:"Mapping where keys are process handler names and the values are a dictionary, where each dictionary can define the ``enabled`` and ``priority`` key, which can be used to toggle the values set on the original process handler declaration."},{name:"max_iterations",required:!1,valid_types:"Int",info:"Maximum number of iterations the work chain will restart the process to finish successfully."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"settings",required:!1,valid_types:"Dict, NoneType",info:"Additional settings."},{name:"shift_energy_windows",required:!1,valid_types:"Bool",info:"If True the `dis_froz_min`, `dis_froz_max`, `dis_win_min`, `dis_win_max` will be shifted by Fermi enerngy. False is the default behaviour of wannier90."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"The ``output_parameters`` output node of the successful calculation."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"interpolated_bands",required:!1,valid_types:"BandsData",info:"The interpolated band structure by Wannier90 (if any)."},{name:"nnkp_file",required:!1,valid_types:"SinglefileData",info:"The ``.nnkp`` file, produced only in -pp (postproc) mode."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:301,message:"The sub process excepted."},{status:302,message:"The sub process was killed."},{status:401,message:"The maximum number of iterations was exceeded."},{status:401,message:"Unrecoverable bvectors error."},{status:402,message:"The process failed for an unknown reason, twice in a row."},{status:402,message:"Unrecoverable disentanglement error."},{status:403,message:"Unrecoverable cube format error."},{status:404,message:"The stdout output file was incomplete probably because the calculation got interrupted."}]},class:"aiida_wannier90_workflows.workflows.base.wannier90:Wannier90BaseWorkChain"},"wannier90_workflows.open_grid":{description:["WorkChain using open_grid.x to bypass the nscf step.",""," The open_grid.x unfolds the symmetrized kmesh to a full kmesh, thus"," the full-kmesh nscf step can be avoided.",""," 2 schemes:"," 1. scf w/ symmetry, more nbnd -> open_grid -> pw2wannier90 -> wannier90"," 2. scf w/ symmetry, default nbnd -> nscf w/ symm, more nbnd -> open_grid"," -> pw2wannier90 -> wannier90"],spec:{inputs:[{name:"pw2wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Pw2wannier90BaseWorkChain`."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Wannier90BaseWorkChain`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If True, work directories of all called calculation will be cleaned at the end of execution."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the NSCF calculation."},{name:"open_grid",required:!1,valid_types:"Data",info:"Inputs for the `OpenGridBaseWorkChain`, if not specified the open_grid step is skipped."},{name:"projwfc",required:!1,valid_types:"Data",info:"Inputs for the `ProjwfcBaseWorkChain`."},{name:"scf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the SCF calculation."}],outputs:[{name:"pw2wannier90",required:!0,valid_types:"",info:""},{name:"wannier90",required:!0,valid_types:"",info:""},{name:"wannier90_pp",required:!0,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"",info:""},{name:"open_grid",required:!1,valid_types:"",info:""},{name:"projwfc",required:!1,valid_types:"",info:""},{name:"scf",required:!1,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"the scf PwBaseWorkChain sub process failed"},{status:430,message:"the nscf PwBaseWorkChain sub process failed"},{status:440,message:"the ProjwfcBaseWorkChain sub process failed"},{status:450,message:"the postproc Wannier90BaseWorkChain sub process failed"},{status:460,message:"the Pw2wannier90BaseWorkChain sub process failed"},{status:470,message:"the Wannier90BaseWorkChain sub process failed"},{status:480,message:"outputs sanity check failed"},{status:490,message:"the OpenGridBaseWorkChain sub process failed"}]},class:"aiida_wannier90_workflows.workflows.open_grid:Wannier90OpenGridWorkChain"},"wannier90_workflows.optimize":{description:["Workchain to optimize dis_proj_min/max for projectability disentanglement."],spec:{inputs:[{name:"pw2wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Pw2wannier90BaseWorkChain`."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Wannier90BaseWorkChain`."},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Explicit kpoints to use for the wannier90 bands interpolation. If specified, wannier90 will use this list of kpoints and will not use the `bands_num_points` mechanism to auto generate a list of kpoints along the kpath. If not specified, the workchain will run seekpath to generate a primitive cell and a bands_kpoints. Specify either this or `bands_kpoints` or `bands_kpoints_distance`. This ensures the wannier interpolated bands has the exact same number of kpoints as PW bands, to calculate bands distance."},{name:"bands_kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"Minimum kpoints distance for seekpath to generate a list of kpoints along the path. Specify either this or `bands_kpoints` or `kpoint_path`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If True, work directories of all called calculation will be cleaned at the end of execution."},{name:"kpoint_path",required:!1,valid_types:"KpointsData, NoneType",info:"High symmetry kpoints to use for the wannier90 bands interpolation. If specified, the high symmetry kpoint labels will be used and wannier90 will use the `bands_num_points` mechanism to auto generate a list of kpoints along the kpath. If not specified, the workchain will run seekpath to generate a primitive cell and a bands_kpoints. Specify either this or `bands_kpoints` or `bands_kpoints_distance`."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the NSCF calculation."},{name:"open_grid",required:!1,valid_types:"Data",info:"Inputs for the `OpenGridBaseWorkChain`, if not specified the open_grid step is skipped."},{name:"optimize_bands_distance_threshold",required:!1,valid_types:"Float, NoneType",info:"If provided, during the iteration of dis_proj_min/max, if the bands distance is smaller than this threshold, the optimization will stop. Unit is eV."},{name:"optimize_disproj",required:!1,valid_types:"Bool",info:"If True iterate dis_proj_min/max to find the best MLWFs for projectability disentanglement."},{name:"optimize_disprojmax_range",required:!1,valid_types:"List",info:"The range to iterate dis_proj_min. `None` means disabling projectability disentanglement."},{name:"optimize_disprojmin_range",required:!1,valid_types:"List",info:"The range to iterate dis_proj_max. `None` means disabling projectability disentanglement."},{name:"optimize_reference_bands",required:!1,valid_types:"BandsData, NoneType",info:"If provided, during the iteration of dis_proj_min/max, the BandsData will be the reference for calculating bands distance, the final optimal MLWFs will be selected based on both spreads and bands distance. If not provided, spreads will be the criterion for selecting optimal MLWFs. The bands distance is calculated for bands below Fermi energy + 2eV."},{name:"optimize_spreads_imbalence_threshold",required:!1,valid_types:"Float, NoneType",info:"If provided, during the iteration of dis_proj_min/max, if the spreads imbalence is smaller than this threshold, the optimization will stop."},{name:"projwfc",required:!1,valid_types:"Data",info:"Inputs for the `ProjwfcBaseWorkChain`."},{name:"scf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the SCF calculation."},{name:"separate_plotting",required:!1,valid_types:"Bool",info:"If True separate the maximal localisation and the plotting of bands/Wannier function in two steps. This allows reusing the chk file to restart plotting if it were crashed due to memory issue."}],outputs:[{name:"band_structure",required:!0,valid_types:"BandsData",info:"The Wannier interpolated band structure."},{name:"pw2wannier90",required:!0,valid_types:"",info:""},{name:"wannier90",required:!0,valid_types:"",info:""},{name:"wannier90_pp",required:!0,valid_types:"",info:""},{name:"bands_distance",required:!1,valid_types:"Float",info:"Bands distances between reference bands and Wannier interpolated bands for Ef to Ef+5eV."},{name:"nscf",required:!1,valid_types:"",info:""},{name:"open_grid",required:!1,valid_types:"",info:""},{name:"primitive_structure",required:!1,valid_types:"StructureData",info:"The normalized and primitivized structure for which the calculations are computed."},{name:"projwfc",required:!1,valid_types:"",info:""},{name:"scf",required:!1,valid_types:"",info:""},{name:"seekpath_parameters",required:!1,valid_types:"Dict",info:"The parameters used in the SeeKpath call to normalize the input or relaxed structure."},{name:"wannier90_optimal",required:!1,valid_types:"",info:""},{name:"wannier90_plot",required:!1,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"the scf PwBaseWorkChain sub process failed"},{status:430,message:"the nscf PwBaseWorkChain sub process failed"},{status:440,message:"the ProjwfcBaseWorkChain sub process failed"},{status:450,message:"the postproc Wannier90BaseWorkChain sub process failed"},{status:460,message:"the Pw2wannier90BaseWorkChain sub process failed"},{status:470,message:"the Wannier90BaseWorkChain sub process failed"},{status:480,message:"outputs sanity check failed"},{status:490,message:"the OpenGridBaseWorkChain sub process failed"},{status:500,message:"the Wannier90Calculation plotting sub process failed"},{status:501,message:"All the trials on dis_proj_min/max have failed, cannot compare spreads"}]},class:"aiida_wannier90_workflows.workflows.optimize:Wannier90OptimizeWorkChain"},"wannier90_workflows.projwfcbands":{description:["WorkChain to compute QE projected band structure for a given structure."],spec:{inputs:[{name:"bands",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the BANDS calculation."},{name:"projwfc",required:!0,valid_types:"Data",info:"Inputs for the `ProjwfcBaseWorkChain` for the projwfc.x calculation."},{name:"scf",required:!0,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the SCF calculation."},{name:"structure",required:!0,valid_types:"StructureData",info:"The inputs structure."},{name:"bands_kpoints",required:!1,valid_types:"KpointsData, NoneType",info:"Explicit kpoints to use for the BANDS calculation. Specify either this or `bands_kpoints_distance`."},{name:"bands_kpoints_distance",required:!1,valid_types:"Float, NoneType",info:"Minimum kpoints distance for the BANDS calculation. Specify either this or `bands_kpoints`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If `True`, work directories of all called calculation will be cleaned at the end of execution."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nbands_factor",required:!1,valid_types:"Float, NoneType",info:"The number of bands for the BANDS calculation is that used for the SCF multiplied by this factor."},{name:"relax",required:!1,valid_types:"Data",info:"Inputs for the `PwRelaxWorkChain`, if not specified at all, the relaxation step is skipped."}],outputs:[{name:"band_parameters",required:!0,valid_types:"Dict",info:"The output parameters of the BANDS `PwBaseWorkChain`."},{name:"band_structure",required:!0,valid_types:"BandsData",info:"The computed band structure."},{name:"projwfc",required:!0,valid_types:"",info:""},{name:"scf_parameters",required:!0,valid_types:"Dict",info:"The output parameters of the SCF `PwBaseWorkChain`."},{name:"primitive_structure",required:!1,valid_types:"StructureData",info:"The normalized and primitivized structure for which the bands are computed."},{name:"seekpath_parameters",required:!1,valid_types:"Dict",info:"The parameters used in the SeeKpath call to normalize the input or relaxed structure."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:201,message:"Cannot specify both `nbands_factor` and `bands.pw.parameters.SYSTEM.nbnd`."},{status:202,message:"Cannot specify both `bands_kpoints` and `bands_kpoints_distance`."},{status:401,message:"The PwRelaxWorkChain sub process failed"},{status:402,message:"The scf PwBasexWorkChain sub process failed"},{status:403,message:"The bands PwBasexWorkChain sub process failed"},{status:404,message:"The ProjwfcBaseWorkChain sub process failed"}]},class:"aiida_wannier90_workflows.workflows.projwfcbands:ProjwfcBandsWorkChain"},"wannier90_workflows.split":{description:["Workchain to split valence+conduction into two calculations for val and cond, respectively."],spec:{inputs:[{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If True, work directories of all called calculation will be cleaned at the end of execution."},{name:"cond",required:!1,valid_types:"Data",info:"Inputs for the `Wannier90BaseWorkChain` for the cond Wannierization."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"split",required:!1,valid_types:"Data",info:"Inputs for the `Wannier90SplitCalculation`."},{name:"val",required:!1,valid_types:"Data",info:"Inputs for the `Wannier90BaseWorkChain` for the val Wannierization."},{name:"valcond",required:!1,valid_types:"Data",info:"Inputs for the `Wannier90OptimizeWorkChain` for the val+cond Wannierization."}],outputs:[{name:"bands_distance",required:!0,valid_types:"",info:""},{name:"cond",required:!1,valid_types:"",info:""},{name:"split",required:!1,valid_types:"",info:""},{name:"val",required:!1,valid_types:"",info:""},{name:"valcond",required:!1,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:500,message:"All the trials on dis_proj_min/max have failed, cannot compare bands distance"},{status:501,message:"the split calculation has failed"},{status:502,message:"the valence Wannier90Calculation has failed"},{status:503,message:"the conduction Wannier90Calculation has failed"},{status:520,message:"All the orbitals are occupied, cannot split into valence and conduction"}]},class:"aiida_wannier90_workflows.workflows.split:Wannier90SplitWorkChain"},"wannier90_workflows.wannier90":{description:["Workchain to obtain maximally localised Wannier functions (MLWF).",""," Run the following steps:"," scf -> nscf -> projwfc -> wannier90 postproc -> pw2wannier90 -> wannier90"],spec:{inputs:[{name:"pw2wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Pw2wannier90BaseWorkChain`."},{name:"structure",required:!0,valid_types:"StructureData",info:"The input structure."},{name:"wannier90",required:!0,valid_types:"Data",info:"Inputs for the `Wannier90BaseWorkChain`."},{name:"clean_workdir",required:!1,valid_types:"Bool",info:"If True, work directories of all called calculation will be cleaned at the end of execution."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the NSCF calculation."},{name:"projwfc",required:!1,valid_types:"Data",info:"Inputs for the `ProjwfcBaseWorkChain`."},{name:"scf",required:!1,valid_types:"Data",info:"Inputs for the `PwBaseWorkChain` for the SCF calculation."}],outputs:[{name:"pw2wannier90",required:!0,valid_types:"",info:""},{name:"wannier90",required:!0,valid_types:"",info:""},{name:"wannier90_pp",required:!0,valid_types:"",info:""},{name:"nscf",required:!1,valid_types:"",info:""},{name:"projwfc",required:!1,valid_types:"",info:""},{name:"scf",required:!1,valid_types:"",info:""}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:420,message:"the scf PwBaseWorkChain sub process failed"},{status:430,message:"the nscf PwBaseWorkChain sub process failed"},{status:440,message:"the ProjwfcBaseWorkChain sub process failed"},{status:450,message:"the postproc Wannier90BaseWorkChain sub process failed"},{status:460,message:"the Pw2wannier90BaseWorkChain sub process failed"},{status:470,message:"the Wannier90BaseWorkChain sub process failed"},{status:480,message:"outputs sanity check failed"}]},class:"aiida_wannier90_workflows.workflows.wannier90:Wannier90WorkChain"}},console_scripts:{"aiida-wannier90-workflows":"aiida_wannier90_workflows.cli:cmd_root"}},commits_count:27,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:10},{colorclass:"purple",text:"Console scripts",count:1}],pip_install_cmd:"pip install aiida-wannier90-workflows",is_installable:"True"},"aiida-wien2k":{code_home:"https://github.com/rubel75/aiida-wien2k",entry_point_prefix:"wien2k",name:"aiida-wien2k",package_name:"aiida_wien2k",hosted_on:"github.com",metadata:{},aiida_version:null,entry_points:{},commits_count:14,development_status:"planning",errors:[],warnings:["E001: Cannot fetch all data from PyPI and missing plugin_info key!","W002: AiiDA version not found"],summaryinfo:[],pip_install_cmd:"See source code repository."},"aiida-yambo":{code_home:"https://github.com/yambo-code/yambo-aiida/",development_status:"stable",entry_point_prefix:"yambo",pip_url:"aiida-yambo",plugin_info:"https://raw.github.com/yambo-code/yambo-aiida/master/setup.json",name:"aiida-yambo",package_name:"aiida_yambo",hosted_on:"github.com",metadata:{release_date:"2020-11-05",description:"YAMBO plugin and workflows for AiiDA",author:"Miki Bonacci, Michael Atambo, Antimo Marrazzo, Prandini Gianluca",author_email:"miki.bonacci@unimore.it",license:"MIT",home_page:"https://github.com/yambo-code/yambo-aiida",classifiers:["Environment :: Plugins","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Topic :: Scientific/Engineering :: Physics"],version:"1.3.0"},aiida_version:">=1.0.0a2",entry_points:{"aiida.calculations":{"yambo.yambo":{description:["AiiDA plugin for the Yambo code."," For more information, refer to http://www.yambo-code.org/"," https://github.com/yambo-code/yambo-aiida and http://aiida-yambo.readthedocs.io/en/latest/"],spec:{inputs:[{name:"code",required:!0,valid_types:"Code",info:"Use a main code for yambo calculation"},{name:"parameters",required:!0,valid_types:"Dict",info:"Use a node that specifies the input parameters"},{name:"parent_folder",required:!0,valid_types:"RemoteData",info:'Use a remote folder as parent folder (for "restarts and similar"'},{name:"settings",required:!0,valid_types:"Dict",info:"Use an additional node for special settings"},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"precode_parameters",required:!1,valid_types:"Dict, NoneType",info:"Use a node that specifies the input parameters for the yambo precode"},{name:"preprocessing_code",required:!1,valid_types:"Code, NoneType",info:"Use a preprocessing code for starting yambo"},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"returns the output parameters"},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"array_alpha",required:!1,valid_types:"ArrayData",info:"returns the alpha array"},{name:"array_alpha_array",required:!1,valid_types:"ArrayData",info:"returns the alpha array"},{name:"array_alpha_bands",required:!1,valid_types:"ArrayData",info:"returns the alpha array bands"},{name:"array_eels",required:!1,valid_types:"ArrayData",info:"returns the eels array"},{name:"array_eps",required:!1,valid_types:"ArrayData",info:"returns the eps array"},{name:"array_ndb",required:!1,valid_types:"ArrayData",info:"returns the array for ndb"},{name:"array_ndb_HFlocXC",required:!1,valid_types:"ArrayData",info:"returns the array ndb for HFlocXC"},{name:"array_ndb_QP",required:!1,valid_types:"ArrayData",info:"returns the array for ndbQP"},{name:"array_qp",required:!1,valid_types:"ArrayData",info:"returns the quasiparticle array band structure"},{name:"bands_quasiparticle",required:!1,valid_types:"BandsData",info:"returns the quasiparticle band structure"},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"},{status:300,message:"The retrieved folder data node could not be accessed."},{status:301,message:"time exceeded the max walltime"},{status:302,message:"failed calculation for some reason: could be a low number of conduction bands"},{status:303,message:"Unexpected behavior of YamboFolder"},{status:304,message:"parallelization error"},{status:305,message:"memory issues"}]},class:"aiida_yambo.calculations.yambo:YamboCalculation"}},"aiida.data":{},"aiida.parsers":{"yambo.yambo":"aiida_yambo.parsers.parsers:YamboParser"}},commits_count:47,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1}],pip_install_cmd:"pip install aiida-yambo",is_installable:"True"},"aiida-yambo-wannier90":{code_home:"https://github.com/aiidaplugins/aiida-yambo-wannier90",entry_point_prefix:"yambo_wannier90",pip_url:"aiida-yambo-wannier90",plugin_info:"https://raw.githubusercontent.com/aiidaplugins/aiida-yambo-wannier90/main/pyproject.toml",documentation_url:"https://aiida-yambo-wannier90.readthedocs.io/en/latest/",version_file:"https://raw.githubusercontent.com/aiidaplugins/aiida-yambo-wannier90/main/aiida_yambo_wannier90/__init__.py",name:"aiida-yambo-wannier90",package_name:"aiida_yambo_wannier90",hosted_on:"github.com",metadata:{release_date:"2022-07-06",description:"Plugin to combine Wannier90 interpolations with GW corrections computed by Yambo",author:"The AiiDA Team",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Natural Language :: English","Programming Language :: Python"],version:"0.1.0b0"},aiida_version:">=1.6.4,<3",entry_points:{"aiida.calculations":{"yambo_wannier90.gw2wannier90":"aiida_yambo_wannier90.calculations.gw2wannier90:Gw2wannier90Calculation"},"aiida.parsers":{"yambo_wannier90.gw2wannier90":"aiida_yambo_wannier90.parsers.gw2wannier90:Gw2wannier90Parser"},"aiida.workflows":{yambo_wannier90:"aiida_yambo_wannier90.workflows:YamboWannier90WorkChain"},console_scripts:{"aiida-yambo-wannier90":"aiida_yambo_wannier90.cli:cmd_root"}},commits_count:21,development_status:"beta",errors:[`E001: Failed to install plugin aiida-yambo-wannier90
Collecting aiida-yambo-wannier90
+  Downloading aiida_yambo_wannier90-0.1.0b0-py3-none-any.whl.metadata (3.4 kB)
+Requirement already satisfied: aiida-core<3,>=1.6.4 in /opt/conda/lib/python3.10/site-packages (from aiida-yambo-wannier90) (2.5.2)
+Collecting voluptuous (from aiida-yambo-wannier90)
+  Downloading voluptuous-0.15.2-py3-none-any.whl.metadata (20 kB)
+Collecting aiida-yambo>=1.3.0 (from aiida-yambo-wannier90)
+  Downloading aiida_yambo-1.3.0-py3-none-any.whl.metadata (1.1 kB)
+Collecting aiida-wannier90-workflows>=1.0.1 (from aiida-yambo-wannier90)
+  Downloading aiida_wannier90_workflows-2.3.0-py3-none-any.whl.metadata (5.6 kB)
+Requirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (1.13.1)
+Requirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.4.2)
+Requirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (6.8.1)
+Requirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.18.0)
+Requirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.1.10)
+Requirement already satisfied: click~=8.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (8.1.7)
+Requirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (1.1.0)
+Requirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.16)
+Requirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.20.3)
+Requirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (8.25.0)
+Requirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.18.2)
+Requirement already satisfied: jinja2~=3.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (3.1.4)
+Requirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.7.8)
+Requirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (6.11.0)
+Requirement already satisfied: numpy~=1.21 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (1.26.4)
+Requirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (2.12.0)
+Requirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.21.10)
+Requirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.2.4)
+Requirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (5.9.8)
+Requirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (2.9.9)
+Requirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (2.7.3)
+Requirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (2021.3)
+Requirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (6.0.1)
+Requirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (2.32.3)
+Requirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (2.0.30)
+Requirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.8.10)
+Requirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (4.66.4)
+Requirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (0.9.5)
+Requirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core<3,>=1.6.4->aiida-yambo-wannier90) (1.16.0)
+Collecting aiida-pseudo>=0.6 (from aiida-wannier90-workflows>=1.0.1->aiida-yambo-wannier90)
+  Downloading aiida_pseudo-1.6.0-py3-none-any.whl.metadata (10 kB)
+Collecting aiida-quantumespresso>=4.4 (from aiida-wannier90-workflows>=1.0.1->aiida-yambo-wannier90)
+  Downloading aiida_quantumespresso-4.6.0-py3-none-any.whl.metadata (35 kB)
+Collecting aiida-wannier90>=2.2 (from aiida-wannier90-workflows>=1.0.1->aiida-yambo-wannier90)
+  Downloading aiida_wannier90-2.2.0-py3-none-any.whl.metadata (4.0 kB)
+Requirement already satisfied: colorama in /opt/conda/lib/python3.10/site-packages (from aiida-wannier90-workflows>=1.0.1->aiida-yambo-wannier90) (0.4.6)
+INFO: pip is looking at multiple versions of aiida-yambo to determine which version is compatible with other requirements. This could take a while.
+Collecting aiida-wannier90-workflows>=1.0.1 (from aiida-yambo-wannier90)
+  Downloading aiida_wannier90_workflows-2.2.0-py3-none-any.whl.metadata (5.4 kB)
+  Downloading aiida_wannier90_workflows-2.1.0-py3-none-any.whl.metadata (4.7 kB)
+  Downloading aiida_wannier90_workflows-1.0.2-py3-none-any.whl.metadata (969 bytes)
+INFO: pip is looking at multiple versions of aiida-wannier90-workflows to determine which version is compatible with other requirements. This could take a while.
+  Downloading aiida_wannier90_workflows-1.0.1-py2.py3-none-any.whl.metadata (925 bytes)
+ERROR: Cannot install aiida-yambo-wannier90 and aiida-yambo-wannier90==0.1.0b0 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-yambo-wannier90 0.1.0b0 depends on aiida-core<3 and >=1.6.4
+    aiida-wannier90-workflows 1.0.2 depends on aiida-core<2 and >=1.1.1
+    aiida-yambo-wannier90 0.1.0b0 depends on aiida-core<3 and >=1.6.4
+    aiida-wannier90-workflows 1.0.1 depends on aiida-core<2 and >=1.1.1
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:1},{colorclass:"purple",text:"Console scripts",count:1}],pip_install_cmd:"pip install --pre aiida-yambo-wannier90",is_installable:"False"},"aiida-yascheduler":{code_home:"https://github.com/tilde-lab/yascheduler",documentation_url:"https://github.com/tilde-lab/yascheduler",entry_point_prefix:"yascheduler",pip_url:"yascheduler",plugin_info:"https://raw.githubusercontent.com/tilde-lab/yascheduler/master/pyproject.toml",name:"aiida-yascheduler",package_name:"aiida_yascheduler",hosted_on:"github.com",metadata:{release_date:"2023-07-29",description:"Yet another computing scheduler and cloud orchestration engine",author:"Andrey Sobolev",author_email:"Evgeny Blokhin , Sergei Korolev ",classifiers:["Development Status :: 4 - Beta","Framework :: AiiDA","Intended Audience :: Science/Research","License :: OSI Approved :: MIT License","Programming Language :: Python :: 3","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.7","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Topic :: Scientific/Engineering :: Chemistry","Topic :: Scientific/Engineering :: Information Analysis","Topic :: Scientific/Engineering :: Physics","Topic :: Software Development :: Libraries :: Python Modules"],version:"1.2.0"},aiida_version:null,entry_points:{"aiida.schedulers":{yascheduler:"yascheduler.aiida_plugin:YaScheduler"},console_scripts:{yainit:"yascheduler.utils:init",yanodes:"yascheduler.utils:show_nodes",yascheduler:"yascheduler.utils:daemonize",yasetnode:"yascheduler.utils:manage_node",yastatus:"yascheduler.utils:check_status",yasubmit:"yascheduler.utils:submit"}},commits_count:8,development_status:"beta",errors:[`E002: Failed to import package aiida_yascheduler
Traceback (most recent call last):
+  File "", line 1, in 
+ModuleNotFoundError: No module named 'aiida_yascheduler'
+
`],warnings:["W002: AiiDA version not found"],summaryinfo:[{colorclass:"purple",text:"Console scripts",count:6},{colorclass:"orange",text:"Other (Schedulers)",count:1}],pip_install_cmd:"pip install yascheduler",is_installable:"True"},"aiida-z2pack":{code_home:"https://github.com/AntimoMarrazzo/aiida-z2pack",entry_point_prefix:"z2pack",pip_url:"git+https://github.com/AntimoMarrazzo/aiida-z2pack",name:"aiida-z2pack",package_name:"aiida_z2pack",hosted_on:"github.com",metadata:{description:"The official AiiDA plugin for z2pack",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11"],author:"Antimo Marrazzo",author_email:"antimo.marrazzo@epfl.ch"},aiida_version:">=2.1.2,<=2.2.2",entry_points:{"aiida.calculations":{"z2pack.z2pack":"aiida_z2pack.calculations.z2pack:Z2packCalculation"},"aiida.parsers":{"z2pack.z2pack":"aiida_z2pack.parsers.z2pack:Z2packParser"},"aiida.workflows":{"z2pack.base":"aiida_z2pack.workchains.base:Z2packBaseWorkChain","z2pack.qsh":"aiida_z2pack.workchains.parity:Z2QSHworkchain","z2pack.3DChern":"aiida_z2pack.workchains.chern:Z2pack3DChernWorkChain","z2pack.refine":"aiida_z2pack.workchains.refine:RefineCrossingsPosition"}},commits_count:0,development_status:"stable",errors:[`E001: Failed to install plugin aiida-z2pack
Collecting git+https://github.com/AntimoMarrazzo/aiida-z2pack
+  Cloning https://github.com/AntimoMarrazzo/aiida-z2pack to /tmp/pip-req-build-4c5_7zw5
+  Running command git clone --filter=blob:none --quiet https://github.com/AntimoMarrazzo/aiida-z2pack /tmp/pip-req-build-4c5_7zw5
+  Resolved https://github.com/AntimoMarrazzo/aiida-z2pack to commit 01a06de726c8b3eab7d18d1a36b4631ad8216f8b
+  Installing build dependencies: started
+  Installing build dependencies: finished with status 'done'
+  Getting requirements to build wheel: started
+  Getting requirements to build wheel: finished with status 'done'
+  Preparing metadata (pyproject.toml): started
+  Preparing metadata (pyproject.toml): finished with status 'done'
+Requirement already satisfied: numpy in /opt/conda/lib/python3.10/site-packages (from aiida-z2pack==3.0.1) (1.26.4)
+Collecting z2pack==2.1.1 (from aiida-z2pack==3.0.1)
+  Downloading z2pack-2.1.1-py3-none-any.whl.metadata (2.3 kB)
+INFO: pip is looking at multiple versions of aiida-z2pack to determine which version is compatible with other requirements. This could take a while.
+ERROR: Cannot install aiida-z2pack==3.0.1 because these package versions have conflicting dependencies.
+
+The conflict is caused by:
+    aiida-z2pack 3.0.1 depends on aiida-core<=2.2.2 and >=2.1.2
+    The user requested (constraint) aiida-core==2.5.2
+
+To fix this you could try to:
+1. loosen the range of package versions you've specified
+2. remove package versions to allow pip attempt to solve the dependency conflict
+
+ERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts
+
`],warnings:[],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"green",text:"Workflows",count:4}],pip_install_cmd:"pip install git+https://github.com/AntimoMarrazzo/aiida-z2pack",is_installable:"False"},"aiida-zeopp":{code_home:"https://github.com/lsmo-epfl/aiida-zeopp",development_status:"stable",entry_point_prefix:"zeopp",pip_url:"aiida-zeopp",plugin_info:"https://raw.github.com/lsmo-epfl/aiida-zeopp/master/setup.json",name:"aiida-zeopp",package_name:"aiida_zeopp",hosted_on:"github.com",metadata:{release_date:"2023-08-26",description:"AiiDA plugin for zeo++",author_email:"Leopold Talirz , Miriam Pougin ",classifiers:["Development Status :: 5 - Production/Stable","Framework :: AiiDA","License :: OSI Approved :: MIT License","Programming Language :: Python","Programming Language :: Python :: 3.10","Programming Language :: Python :: 3.11","Programming Language :: Python :: 3.8","Programming Language :: Python :: 3.9"],version:"2.0.0"},aiida_version:">=2.3,<3.0",entry_points:{"aiida.calculations":{"zeopp.network":{description:["AiiDA calculation plugin for the zeo++ network binary"],spec:{inputs:[{name:"parameters",required:!0,valid_types:"NetworkParameters",info:"command line parameters for zeo++"},{name:"structure",required:!0,valid_types:"CifData",info:"input structure to be analyzed"},{name:"atomic_radii",required:!1,valid_types:"SinglefileData, NoneType",info:"atomic radii file"},{name:"code",required:!1,valid_types:"AbstractCode, NoneType",info:"The `Code` to use for this job. This input is required, unless the `remote_folder` input is specified, which means an existing job is being imported and no code will actually be run."},{name:"metadata",required:!1,valid_types:"",info:""},{name:"monitors",required:!1,valid_types:"Dict",info:"Add monitoring functions that can inspect output files while the job is running and decide to prematurely terminate the job."},{name:"remote_folder",required:!1,valid_types:"RemoteData, NoneType",info:"Remote directory containing the results of an already completed calculation job without AiiDA. The inputs should be passed to the `CalcJob` as normal but instead of launching the actual job, the engine will recreate the input files and then proceed straight to the retrieve step where the files of this `RemoteData` will be retrieved as if it had been actually launched through AiiDA. If a parser is defined in the inputs, the results are parsed and attached as output nodes as usual."}],outputs:[{name:"output_parameters",required:!0,valid_types:"Dict",info:"key-value pairs parsed from zeo++ output file(s)."},{name:"remote_folder",required:!0,valid_types:"RemoteData",info:"Input files necessary to run the process will be stored in this folder node."},{name:"retrieved",required:!0,valid_types:"FolderData",info:"Files that are retrieved by the daemon will be stored in this node. By default the stdout and stderr of the scheduler will be added, but one can add more by specifying them in `CalcInfo.retrieve_list`."},{name:"block",required:!1,valid_types:"SinglefileData",info:"Blocked pockets fileoutput file."},{name:"remote_stash",required:!1,valid_types:"RemoteStashData",info:"Contents of the `stash.source_list` option are stored in this remote folder after job completion."}],exit_codes:[{status:0,message:"Calculation completed successfully."},{status:1,message:"The process has failed with an unspecified error."},{status:2,message:"The process failed with legacy failure mode."},{status:10,message:"The process returned an invalid output."},{status:11,message:"The process did not register a required output."},{status:100,message:"The process did not have the required `retrieved` output."},{status:101,message:"Not all expected output files were found."},{status:102,message:"Empty block file. This indicates the calculation of blocked pockets did not finish."},{status:110,message:"The job ran out of memory."},{status:120,message:"The job ran out of walltime."},{status:131,message:"The specified account is invalid."},{status:140,message:"The node running the job failed."},{status:150,message:"{message}"}]},class:"aiida_zeopp.calculations.network:NetworkCalculation"}},"aiida.data":{"zeopp.parameters":"aiida_zeopp.data.parameters:NetworkParameters"},"aiida.parsers":{"zeopp.network":"aiida_zeopp.parsers.network:NetworkParser"}},commits_count:0,errors:[],warnings:["W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead."],summaryinfo:[{colorclass:"blue",text:"Calculations",count:1},{colorclass:"brown",text:"Parsers",count:1},{colorclass:"red",text:"Data",count:1}],pip_install_cmd:"pip install aiida-zeopp",is_installable:"True"}},jT=[{name:"Calculations",colorclass:"blue",num_entries:67,total_num:165},{name:"Parsers",colorclass:"brown",num_entries:68,total_num:146},{name:"Data",colorclass:"red",num_entries:38,total_num:122},{name:"Workflows",colorclass:"green",num_entries:50,total_num:215},{name:"Console scripts",colorclass:"purple",num_entries:17,total_num:33},{name:"Other",tooltip:"Aenet potentials, Aiida python serializers, Aiidalab qe properties, ...",colorclass:"orange",num_entries:33,total_num:110}],ET={planning:["Not yet ready to use. Developers welcome!","status-planning-d9644d.svg"],"pre-alpha":["Not yet ready to use. Developers welcome!","status-planning-d9644d.svg"],alpha:["Adds new functionality, not yet ready for production. Testing welcome!","status-alpha-d6af23.svg"],beta:["Adds new functionality, not yet ready for production. Testing welcome!","status-beta-d6af23.svg"],stable:["Ready for production calculations. Bug reports welcome!","status-stable-4cc61e.svg"],mature:["Ready for production calculations. Bug reports welcome!","status-stable-4cc61e.svg"],inactive:["No longer maintained.","status-inactive-bbbbbb.svg"]},FT={"aiida.calculations":"CalcJobs and calculation functions","aiida.parsers":"CalcJob parsers","aiida.data":"Data node types","aiida.cmdline.data":"verdi data commands","aiida.groups":"Group types","aiida.workflows":"WorkChains and work functions","aiida.schedulers":"Job scheduler support","aiida.transports":"Data transport protocols","aiida.tests":"Development test modules","aiida.tools.dbexporters":"Support for exporting to external databases","aiida.tools.dbimporters":"Support for importing from external databases",console_scripts:"Console scripts"},ya={plugins:NT,globalsummary:jT,status_dict:ET,entrypointtypes:FT};var mp={},R_={exports:{}};(function(e){function t(a){return a&&a.__esModule?a:{default:a}}e.exports=t,e.exports.__esModule=!0,e.exports.default=e.exports})(R_);var Zr=R_.exports,Cd={};function q(){return q=Object.assign?Object.assign.bind():function(e){for(var t=1;t{t[a]=I_(e[a])}),t}function xt(e,t,a={clone:!0}){const i=a.clone?q({},e):e;return Xa(e)&&Xa(t)&&Object.keys(t).forEach(s=>{Xa(t[s])&&Object.prototype.hasOwnProperty.call(e,s)&&Xa(e[s])?i[s]=xt(e[s],t[s],a):a.clone?i[s]=Xa(t[s])?I_(t[s]):t[s]:i[s]=t[s]}),i}const BT=Object.freeze(Object.defineProperty({__proto__:null,default:xt,isPlainObject:Xa},Symbol.toStringTag,{value:"Module"}));function Oi(e){let t="https://mui.com/production-error/?code="+e;for(let a=1;aa==null?t:function(...s){t.apply(this,s),a.apply(this,s)},()=>{})}function hl(e,t=166){let a;function i(...s){const r=()=>{e.apply(this,s)};clearTimeout(a),a=setTimeout(r,t)}return i.clear=()=>{clearTimeout(a)},i}function KT(e,t){return()=>null}function Qn(e,t){var a,i;return b.isValidElement(e)&&t.indexOf((a=e.type.muiName)!=null?a:(i=e.type)==null||(i=i._payload)==null||(i=i.value)==null?void 0:i.muiName)!==-1}function _t(e){return e&&e.ownerDocument||document}function ha(e){return _t(e).defaultView||window}function VT(e,t){return()=>null}function Io(e,t){typeof e=="function"?e(t):e&&(e.current=t)}const HT=typeof window<"u"?b.useLayoutEffect:b.useEffect,Ia=HT;let Im=0;function GT(e){const[t,a]=b.useState(e),i=e||t;return b.useEffect(()=>{t==null&&(Im+=1,a(`mui-${Im}`))},[t]),i}const Nm=so["useId".toString()];function yp(e){if(Nm!==void 0){const t=Nm();return e??t}return GT(e)}function XT(e,t,a,i,s){return null}function Uu({controlled:e,default:t,name:a,state:i="value"}){const{current:s}=b.useRef(e!==void 0),[r,n]=b.useState(t),o=s?e:r,l=b.useCallback(d=>{s||n(d)},[]);return[o,l]}function Si(e){const t=b.useRef(e);return Ia(()=>{t.current=e}),b.useRef((...a)=>(0,t.current)(...a)).current}function Ke(...e){return b.useMemo(()=>e.every(t=>t==null)?null:t=>{e.forEach(a=>{Io(a,t)})},e)}const jm={};function JT(e,t){const a=b.useRef(jm);return a.current===jm&&(a.current=e(t)),a}const QT=[];function YT(e){b.useEffect(e,QT)}class gl{constructor(){this.currentId=null,this.clear=()=>{this.currentId!==null&&(clearTimeout(this.currentId),this.currentId=null)},this.disposeEffect=()=>this.clear}static create(){return new gl}start(t,a){this.clear(),this.currentId=setTimeout(()=>{this.currentId=null,a()},t)}}function B_(){const e=JT(gl.create).current;return YT(e.disposeEffect),e}let yl=!0,Ku=!1;const ZT=new gl,ek={text:!0,search:!0,url:!0,tel:!0,email:!0,password:!0,number:!0,date:!0,month:!0,week:!0,time:!0,datetime:!0,"datetime-local":!0};function tk(e){const{type:t,tagName:a}=e;return!!(a==="INPUT"&&ek[t]&&!e.readOnly||a==="TEXTAREA"&&!e.readOnly||e.isContentEditable)}function ak(e){e.metaKey||e.altKey||e.ctrlKey||(yl=!0)}function Sd(){yl=!1}function ik(){this.visibilityState==="hidden"&&Ku&&(yl=!0)}function sk(e){e.addEventListener("keydown",ak,!0),e.addEventListener("mousedown",Sd,!0),e.addEventListener("pointerdown",Sd,!0),e.addEventListener("touchstart",Sd,!0),e.addEventListener("visibilitychange",ik,!0)}function rk(e){const{target:t}=e;try{return t.matches(":focus-visible")}catch{}return yl||tk(t)}function O_(){const e=b.useCallback(s=>{s!=null&&sk(s.ownerDocument)},[]),t=b.useRef(!1);function a(){return t.current?(Ku=!0,ZT.start(100,()=>{Ku=!1}),t.current=!1,!0):!1}function i(s){return rk(s)?(t.current=!0,!0):!1}return{isFocusVisibleRef:t,onFocus:i,onBlur:a,ref:e}}function M_(e){const t=e.documentElement.clientWidth;return Math.abs(window.innerWidth-t)}function Vu(e,t){const a=q({},t);return Object.keys(e).forEach(i=>{if(i.toString().match(/^(components|slots)$/))a[i]=q({},e[i],a[i]);else if(i.toString().match(/^(componentsProps|slotProps)$/)){const s=e[i]||{},r=t[i];a[i]={},!r||!Object.keys(r)?a[i]=s:!s||!Object.keys(s)?a[i]=r:(a[i]=q({},r),Object.keys(s).forEach(n=>{a[i][n]=Vu(s[n],r[n])}))}else a[i]===void 0&&(a[i]=e[i])}),a}function fe(e,t,a=void 0){const i={};return Object.keys(e).forEach(s=>{i[s]=e[s].reduce((r,n)=>{if(n){const o=t(n);o!==""&&r.push(o),a&&a[n]&&r.push(a[n])}return r},[]).join(" ")}),i}const Em=e=>e,nk=()=>{let e=Em;return{configure(t){e=t},generate(t){return e(t)},reset(){e=Em}}},ok=nk(),_p=ok,lk={active:"active",checked:"checked",completed:"completed",disabled:"disabled",error:"error",expanded:"expanded",focused:"focused",focusVisible:"focusVisible",open:"open",readOnly:"readOnly",required:"required",selected:"selected"};function pe(e,t,a="Mui"){const i=lk[t];return i?`${a}-${i}`:`${_p.generate(e)}-${t}`}function re(e,t,a="Mui"){const i={};return t.forEach(s=>{i[s]=pe(e,s,a)}),i}function dk(e,t=Number.MIN_SAFE_INTEGER,a=Number.MAX_SAFE_INTEGER){return Math.max(t,Math.min(e,a))}const uk=Object.freeze(Object.defineProperty({__proto__:null,default:dk},Symbol.toStringTag,{value:"Module"}));function U(e,t){if(e==null)return{};var a={};for(var i in e)if({}.hasOwnProperty.call(e,i)){if(t.includes(i))continue;a[i]=e[i]}return a}function No(e){return typeof e=="string"}function L_(e,t,a){return e===void 0||No(e)?t:q({},t,{ownerState:q({},t.ownerState,a)})}function W_(e){var t,a,i="";if(typeof e=="string"||typeof e=="number")i+=e;else if(typeof e=="object")if(Array.isArray(e)){var s=e.length;for(t=0;ti.match(/^on[A-Z]/)&&typeof e[i]=="function"&&!t.includes(i)).forEach(i=>{a[i]=e[i]}),a}function Fm(e){if(e===void 0)return{};const t={};return Object.keys(e).filter(a=>!(a.match(/^on[A-Z]/)&&typeof e[a]=="function")).forEach(a=>{t[a]=e[a]}),t}function $_(e){const{getSlotProps:t,additionalProps:a,externalSlotProps:i,externalForwardedProps:s,className:r}=e;if(!t){const v=H(a==null?void 0:a.className,r,s==null?void 0:s.className,i==null?void 0:i.className),f=q({},a==null?void 0:a.style,s==null?void 0:s.style,i==null?void 0:i.style),h=q({},a,s,i);return v.length>0&&(h.className=v),Object.keys(f).length>0&&(h.style=f),{props:h,internalRef:void 0}}const n=z_(q({},s,i)),o=Fm(i),l=Fm(s),d=t(n),c=H(d==null?void 0:d.className,a==null?void 0:a.className,r,s==null?void 0:s.className,i==null?void 0:i.className),p=q({},d==null?void 0:d.style,a==null?void 0:a.style,s==null?void 0:s.style,i==null?void 0:i.style),u=q({},d,a,l,o);return c.length>0&&(u.className=c),Object.keys(p).length>0&&(u.style=p),{props:u,internalRef:d.ref}}function U_(e,t,a){return typeof e=="function"?e(t,a):e}const ck=["elementType","externalSlotProps","ownerState","skipResolvingSlotProps"];function Cs(e){var t;const{elementType:a,externalSlotProps:i,ownerState:s,skipResolvingSlotProps:r=!1}=e,n=U(e,ck),o=r?{}:U_(i,s),{props:l,internalRef:d}=$_(q({},n,{externalSlotProps:o})),c=Ke(d,o==null?void 0:o.ref,(t=e.additionalProps)==null?void 0:t.ref);return L_(a,q({},l,{ref:c}),s)}const pk=b.createContext(void 0);function fk(e){const{theme:t,name:a,props:i}=e;if(!t||!t.components||!t.components[a])return i;const s=t.components[a];return s.defaultProps?Vu(s.defaultProps,i):!s.styleOverrides&&!s.variants?Vu(s,i):i}function mk({props:e,name:t}){const a=b.useContext(pk);return fk({props:e,name:t,theme:{components:a}})}function me(e){return mk(e)}var en={},Pd={exports:{}},Bm;function K_(){return Bm||(Bm=1,function(e){function t(){return e.exports=t=Object.assign?Object.assign.bind():function(a){for(var i=1;i0?Je(Fs,--bt):0,Ss--,Oe===10&&(Ss=1,vl--),Oe}function Ct(){return Oe=bt2||$r(Oe)>3?"":" "}function Ik(e,t){for(;--t&&Ct()&&!(Oe<48||Oe>102||Oe>57&&Oe<65||Oe>70&&Oe<97););return tn(e,Yn()+(t<6&&ma()==32&&Ct()==32))}function Gu(e){for(;Ct();)switch(Oe){case e:return bt;case 34:case 39:e!==34&&e!==39&&Gu(Oe);break;case 40:e===41&&Gu(e);break;case 92:Ct();break}return bt}function Nk(e,t){for(;Ct()&&e+Oe!==47+10;)if(e+Oe===42+42&&ma()===47)break;return"/*"+tn(t,bt-1)+"*"+_l(e===47?e:Ct())}function jk(e){for(;!$r(ma());)Ct();return tn(e,bt)}function Ek(e){return Y_(eo("",null,null,null,[""],e=Q_(e),0,[0],e))}function eo(e,t,a,i,s,r,n,o,l){for(var d=0,c=0,p=n,u=0,v=0,f=0,h=1,T=1,y=1,g=0,_="",w=s,D=r,x=i,k=_;T;)switch(f=g,g=Ct()){case 40:if(f!=108&&Je(k,p-1)==58){Hu(k+=se(Zn(g),"&","&\f"),"&\f")!=-1&&(y=-1);break}case 34:case 39:case 91:k+=Zn(g);break;case 9:case 10:case 13:case 32:k+=Rk(f);break;case 92:k+=Ik(Yn()-1,7);continue;case 47:switch(ma()){case 42:case 47:Pn(Fk(Nk(Ct(),Yn()),t,a),l);break;default:k+="/"}break;case 123*h:o[d++]=la(k)*y;case 125*h:case 59:case 0:switch(g){case 0:case 125:T=0;case 59+c:y==-1&&(k=se(k,/\f/g,"")),v>0&&la(k)-p&&Pn(v>32?Lm(k+";",i,a,p-1):Lm(se(k," ","")+";",i,a,p-2),l);break;case 59:k+=";";default:if(Pn(x=Mm(k,t,a,d,c,s,o,_,w=[],D=[],p),r),g===123)if(c===0)eo(k,t,x,x,w,r,p,o,D);else switch(u===99&&Je(k,3)===110?100:u){case 100:case 108:case 109:case 115:eo(e,x,x,i&&Pn(Mm(e,x,x,0,0,s,o,_,s,w=[],p),D),s,D,p,o,i?w:D);break;default:eo(k,x,x,x,[""],D,0,o,D)}}d=c=v=0,h=y=1,_=k="",p=n;break;case 58:p=1+la(k),v=f;default:if(h<1){if(g==123)--h;else if(g==125&&h++==0&&Ak()==125)continue}switch(k+=_l(g),g*h){case 38:y=c>0?1:(k+="\f",-1);break;case 44:o[d++]=(la(k)-1)*y,y=1;break;case 64:ma()===45&&(k+=Zn(Ct())),u=ma(),c=p=la(_=k+=jk(Yn())),g++;break;case 45:f===45&&la(k)==2&&(h=0)}}return r}function Mm(e,t,a,i,s,r,n,o,l,d,c){for(var p=s-1,u=s===0?r:[""],v=wp(u),f=0,h=0,T=0;f0?u[y]+" "+g:se(g,/&\f/g,u[y])))&&(l[T++]=_);return bl(e,t,a,s===0?vp:o,l,d,c)}function Fk(e,t,a){return bl(e,t,a,H_,_l(Pk()),zr(e,2,-2),0)}function Lm(e,t,a,i){return bl(e,t,a,bp,zr(e,0,i),zr(e,i+1,-1),i)}function _s(e,t){for(var a="",i=wp(e),s=0;s6)switch(Je(e,t+1)){case 109:if(Je(e,t+4)!==45)break;case 102:return se(e,/(.+:)(.+)-([^]+)/,"$1"+ie+"$2-$3$1"+jo+(Je(e,t+3)==108?"$3":"$2-$3"))+e;case 115:return~Hu(e,"stretch")?Z_(se(e,"stretch","fill-available"),t)+e:e}break;case 4949:if(Je(e,t+1)!==115)break;case 6444:switch(Je(e,la(e)-3-(~Hu(e,"!important")&&10))){case 107:return se(e,":",":"+ie)+e;case 101:return se(e,/(.+:)([^;!]+)(;|!.+)?/,"$1"+ie+(Je(e,14)===45?"inline-":"")+"box$3$1"+ie+"$2$3$1"+at+"$2box$3")+e}break;case 5936:switch(Je(e,t+11)){case 114:return ie+e+at+se(e,/[svh]\w+-[tblr]{2}/,"tb")+e;case 108:return ie+e+at+se(e,/[svh]\w+-[tblr]{2}/,"tb-rl")+e;case 45:return ie+e+at+se(e,/[svh]\w+-[tblr]{2}/,"lr")+e}return ie+e+at+e+e}return e}var Kk=function(t,a,i,s){if(t.length>-1&&!t.return)switch(t.type){case bp:t.return=Z_(t.value,t.length);break;case G_:return _s([Js(t,{value:se(t.value,"@","@"+ie)})],s);case vp:if(t.length)return Sk(t.props,function(r){switch(Ck(r,/(::plac\w+|:read-\w+)/)){case":read-only":case":read-write":return _s([Js(t,{props:[se(r,/:(read-\w+)/,":"+jo+"$1")]})],s);case"::placeholder":return _s([Js(t,{props:[se(r,/:(plac\w+)/,":"+ie+"input-$1")]}),Js(t,{props:[se(r,/:(plac\w+)/,":"+jo+"$1")]}),Js(t,{props:[se(r,/:(plac\w+)/,at+"input-$1")]})],s)}return""})}},Vk=[Kk],ev=function(t){var a=t.key;if(a==="css"){var i=document.querySelectorAll("style[data-emotion]:not([data-s])");Array.prototype.forEach.call(i,function(h){var T=h.getAttribute("data-emotion");T.indexOf(" ")!==-1&&(document.head.appendChild(h),h.setAttribute("data-s",""))})}var s=t.stylisPlugins||Vk,r={},n,o=[];n=t.container||document.head,Array.prototype.forEach.call(document.querySelectorAll('style[data-emotion^="'+a+' "]'),function(h){for(var T=h.getAttribute("data-emotion").split(" "),y=1;y=4;++i,s-=4)a=e.charCodeAt(i)&255|(e.charCodeAt(++i)&255)<<8|(e.charCodeAt(++i)&255)<<16|(e.charCodeAt(++i)&255)<<24,a=(a&65535)*1540483477+((a>>>16)*59797<<16),a^=a>>>24,t=(a&65535)*1540483477+((a>>>16)*59797<<16)^(t&65535)*1540483477+((t>>>16)*59797<<16);switch(s){case 3:t^=(e.charCodeAt(i+2)&255)<<16;case 2:t^=(e.charCodeAt(i+1)&255)<<8;case 1:t^=e.charCodeAt(i)&255,t=(t&65535)*1540483477+((t>>>16)*59797<<16)}return t^=t>>>13,t=(t&65535)*1540483477+((t>>>16)*59797<<16),((t^t>>>15)>>>0).toString(36)}var sq={animationIterationCount:1,aspectRatio:1,borderImageOutset:1,borderImageSlice:1,borderImageWidth:1,boxFlex:1,boxFlexGroup:1,boxOrdinalGroup:1,columnCount:1,columns:1,flex:1,flexGrow:1,flexPositive:1,flexShrink:1,flexNegative:1,flexOrder:1,gridRow:1,gridRowEnd:1,gridRowSpan:1,gridRowStart:1,gridColumn:1,gridColumnEnd:1,gridColumnSpan:1,gridColumnStart:1,msGridRow:1,msGridRowSpan:1,msGridColumn:1,msGridColumnSpan:1,fontWeight:1,lineHeight:1,opacity:1,order:1,orphans:1,scale:1,tabSize:1,widows:1,zIndex:1,zoom:1,WebkitLineClamp:1,fillOpacity:1,floodOpacity:1,stopOpacity:1,strokeDasharray:1,strokeDashoffset:1,strokeMiterlimit:1,strokeOpacity:1,strokeWidth:1},rq=!1,nq=/[A-Z]|^ms/g,oq=/_EMO_([^_]+?)_([^]*?)_EMO_/g,ov=function(t){return t.charCodeAt(1)===45},zm=function(t){return t!=null&&typeof t!="boolean"},Rd=V_(function(e){return ov(e)?e:e.replace(nq,"-$&").toLowerCase()}),$m=function(t,a){switch(t){case"animation":case"animationName":if(typeof a=="string")return a.replace(oq,function(i,s,r){return da={name:s,styles:r,next:da},s})}return sq[t]!==1&&!ov(t)&&typeof a=="number"&&a!==0?a+"px":a},lq="Component selectors can only be used in conjunction with @emotion/babel-plugin, the swc Emotion plugin, or another Emotion-aware compiler transform.";function Ur(e,t,a){if(a==null)return"";var i=a;if(i.__emotion_styles!==void 0)return i;switch(typeof a){case"boolean":return"";case"object":{var s=a;if(s.anim===1)return da={name:s.name,styles:s.styles,next:da},s.name;var r=a;if(r.styles!==void 0){var n=r.next;if(n!==void 0)for(;n!==void 0;)da={name:n.name,styles:n.styles,next:da},n=n.next;var o=r.styles+";";return o}return dq(e,t,a)}case"function":{if(e!==void 0){var l=da,d=a(e);return da=l,Ur(e,t,d)}break}}var c=a;if(t==null)return c;var p=t[c];return p!==void 0?p:c}function dq(e,t,a){var i="";if(Array.isArray(a))for(var s=0;s96?mq:hq},Hm=function(t,a,i){var s;if(a){var r=a.shouldForwardProp;s=t.__emotion_forwardProp&&r?function(n){return t.__emotion_forwardProp(n)&&r(n)}:r}return typeof s!="function"&&i&&(s=t.__emotion_forwardProp),s},gq=!1,yq=function(t){var a=t.cache,i=t.serialized,s=t.isStringTag;return rv(a,i,s),cq(function(){return nv(a,i,s)}),null},_q=function e(t,a){var i=t.__emotion_real===t,s=i&&t.__emotion_base||t,r,n;a!==void 0&&(r=a.label,n=a.target);var o=Hm(t,a,i),l=o||Vm(s),d=!l("as");return function(){var c=arguments,p=i&&t.__emotion_styles!==void 0?t.__emotion_styles.slice(0):[];if(r!==void 0&&p.push("label:"+r+";"),c[0]==null||c[0].raw===void 0)p.push.apply(p,c);else{p.push(c[0][0]);for(var u=c.length,v=1;vt(wq(s)?a:s):t;return m.jsx(fq,{styles:i})}/** + * @mui/styled-engine v5.16.6 + * + * @license MIT + * This source code is licensed under the MIT license found in the + * LICENSE file in the root directory of this source tree. + */function fv(e,t){return Xu(e,t)}const Tq=(e,t)=>{Array.isArray(e.__emotion_styles)&&(e.__emotion_styles=t(e.__emotion_styles))},kq=Object.freeze(Object.defineProperty({__proto__:null,GlobalStyles:pv,StyledEngineProvider:bq,ThemeContext:Rl,css:cv,default:fv,internal_processStyles:Tq,keyframes:Il},Symbol.toStringTag,{value:"Module"})),qq=Fa(kq),Dq=Fa(BT),xq=Fa(UT),Cq=Fa($T),Sq=["values","unit","step"],Pq=e=>{const t=Object.keys(e).map(a=>({key:a,val:e[a]}))||[];return t.sort((a,i)=>a.val-i.val),t.reduce((a,i)=>q({},a,{[i.key]:i.val}),{})};function mv(e){const{values:t={xs:0,sm:600,md:900,lg:1200,xl:1536},unit:a="px",step:i=5}=e,s=U(e,Sq),r=Pq(t),n=Object.keys(r);function o(u){return`@media (min-width:${typeof t[u]=="number"?t[u]:u}${a})`}function l(u){return`@media (max-width:${(typeof t[u]=="number"?t[u]:u)-i/100}${a})`}function d(u,v){const f=n.indexOf(v);return`@media (min-width:${typeof t[u]=="number"?t[u]:u}${a}) and (max-width:${(f!==-1&&typeof t[n[f]]=="number"?t[n[f]]:v)-i/100}${a})`}function c(u){return n.indexOf(u)+1`@media (min-width:${xp[e]}px)`};function Na(e,t,a){const i=e.theme||{};if(Array.isArray(t)){const r=i.breakpoints||Gm;return t.reduce((n,o,l)=>(n[r.up(r.keys[l])]=a(t[l]),n),{})}if(typeof t=="object"){const r=i.breakpoints||Gm;return Object.keys(t).reduce((n,o)=>{if(Object.keys(r.values||xp).indexOf(o)!==-1){const l=r.up(o);n[l]=a(t[o],o)}else{const l=o;n[l]=t[l]}return n},{})}return a(t)}function Iq(e={}){var t;return((t=e.keys)==null?void 0:t.reduce((i,s)=>{const r=e.up(s);return i[r]={},i},{}))||{}}function Nq(e,t){return e.reduce((a,i)=>{const s=a[i];return(!s||Object.keys(s).length===0)&&delete a[i],a},t)}function Nl(e,t,a=!0){if(!t||typeof t!="string")return null;if(e&&e.vars&&a){const i=`vars.${t}`.split(".").reduce((s,r)=>s&&s[r]?s[r]:null,e);if(i!=null)return i}return t.split(".").reduce((i,s)=>i&&i[s]!=null?i[s]:null,e)}function Eo(e,t,a,i=a){let s;return typeof e=="function"?s=e(a):Array.isArray(e)?s=e[a]||i:s=Nl(e,a)||i,t&&(s=t(s,i,e)),s}function Fe(e){const{prop:t,cssProperty:a=e.prop,themeKey:i,transform:s}=e,r=n=>{if(n[t]==null)return null;const o=n[t],l=n.theme,d=Nl(l,i)||{};return Na(n,o,p=>{let u=Eo(d,s,p);return p===u&&typeof p=="string"&&(u=Eo(d,s,`${t}${p==="default"?"":X(p)}`,p)),a===!1?u:{[a]:u}})};return r.propTypes={},r.filterProps=[t],r}function jq(e){const t={};return a=>(t[a]===void 0&&(t[a]=e(a)),t[a])}const Eq={m:"margin",p:"padding"},Fq={t:"Top",r:"Right",b:"Bottom",l:"Left",x:["Left","Right"],y:["Top","Bottom"]},Xm={marginX:"mx",marginY:"my",paddingX:"px",paddingY:"py"},Bq=jq(e=>{if(e.length>2)if(Xm[e])e=Xm[e];else return[e];const[t,a]=e.split(""),i=Eq[t],s=Fq[a]||"";return Array.isArray(s)?s.map(r=>i+r):[i+s]}),Cp=["m","mt","mr","mb","ml","mx","my","margin","marginTop","marginRight","marginBottom","marginLeft","marginX","marginY","marginInline","marginInlineStart","marginInlineEnd","marginBlock","marginBlockStart","marginBlockEnd"],Sp=["p","pt","pr","pb","pl","px","py","padding","paddingTop","paddingRight","paddingBottom","paddingLeft","paddingX","paddingY","paddingInline","paddingInlineStart","paddingInlineEnd","paddingBlock","paddingBlockStart","paddingBlockEnd"];[...Cp,...Sp];function an(e,t,a,i){var s;const r=(s=Nl(e,t,!1))!=null?s:a;return typeof r=="number"?n=>typeof n=="string"?n:r*n:Array.isArray(r)?n=>typeof n=="string"?n:r[n]:typeof r=="function"?r:()=>{}}function hv(e){return an(e,"spacing",8)}function sn(e,t){if(typeof t=="string"||t==null)return t;const a=Math.abs(t),i=e(a);return t>=0?i:typeof i=="number"?-i:`-${i}`}function Oq(e,t){return a=>e.reduce((i,s)=>(i[s]=sn(t,a),i),{})}function Mq(e,t,a,i){if(t.indexOf(a)===-1)return null;const s=Bq(a),r=Oq(s,i),n=e[a];return Na(e,n,r)}function gv(e,t){const a=hv(e.theme);return Object.keys(e).map(i=>Mq(e,t,i,a)).reduce(vr,{})}function Ae(e){return gv(e,Cp)}Ae.propTypes={};Ae.filterProps=Cp;function Re(e){return gv(e,Sp)}Re.propTypes={};Re.filterProps=Sp;function Lq(e=8){if(e.mui)return e;const t=hv({spacing:e}),a=(...i)=>(i.length===0?[1]:i).map(r=>{const n=t(r);return typeof n=="number"?`${n}px`:n}).join(" ");return a.mui=!0,a}function jl(...e){const t=e.reduce((i,s)=>(s.filterProps.forEach(r=>{i[r]=s}),i),{}),a=i=>Object.keys(i).reduce((s,r)=>t[r]?vr(s,t[r](i)):s,{});return a.propTypes={},a.filterProps=e.reduce((i,s)=>i.concat(s.filterProps),[]),a}function Wt(e){return typeof e!="number"?e:`${e}px solid`}function Gt(e,t){return Fe({prop:e,themeKey:"borders",transform:t})}const Wq=Gt("border",Wt),zq=Gt("borderTop",Wt),$q=Gt("borderRight",Wt),Uq=Gt("borderBottom",Wt),Kq=Gt("borderLeft",Wt),Vq=Gt("borderColor"),Hq=Gt("borderTopColor"),Gq=Gt("borderRightColor"),Xq=Gt("borderBottomColor"),Jq=Gt("borderLeftColor"),Qq=Gt("outline",Wt),Yq=Gt("outlineColor"),El=e=>{if(e.borderRadius!==void 0&&e.borderRadius!==null){const t=an(e.theme,"shape.borderRadius",4),a=i=>({borderRadius:sn(t,i)});return Na(e,e.borderRadius,a)}return null};El.propTypes={};El.filterProps=["borderRadius"];jl(Wq,zq,$q,Uq,Kq,Vq,Hq,Gq,Xq,Jq,El,Qq,Yq);const Fl=e=>{if(e.gap!==void 0&&e.gap!==null){const t=an(e.theme,"spacing",8),a=i=>({gap:sn(t,i)});return Na(e,e.gap,a)}return null};Fl.propTypes={};Fl.filterProps=["gap"];const Bl=e=>{if(e.columnGap!==void 0&&e.columnGap!==null){const t=an(e.theme,"spacing",8),a=i=>({columnGap:sn(t,i)});return Na(e,e.columnGap,a)}return null};Bl.propTypes={};Bl.filterProps=["columnGap"];const Ol=e=>{if(e.rowGap!==void 0&&e.rowGap!==null){const t=an(e.theme,"spacing",8),a=i=>({rowGap:sn(t,i)});return Na(e,e.rowGap,a)}return null};Ol.propTypes={};Ol.filterProps=["rowGap"];const Zq=Fe({prop:"gridColumn"}),eD=Fe({prop:"gridRow"}),tD=Fe({prop:"gridAutoFlow"}),aD=Fe({prop:"gridAutoColumns"}),iD=Fe({prop:"gridAutoRows"}),sD=Fe({prop:"gridTemplateColumns"}),rD=Fe({prop:"gridTemplateRows"}),nD=Fe({prop:"gridTemplateAreas"}),oD=Fe({prop:"gridArea"});jl(Fl,Bl,Ol,Zq,eD,tD,aD,iD,sD,rD,nD,oD);function vs(e,t){return t==="grey"?t:e}const lD=Fe({prop:"color",themeKey:"palette",transform:vs}),dD=Fe({prop:"bgcolor",cssProperty:"backgroundColor",themeKey:"palette",transform:vs}),uD=Fe({prop:"backgroundColor",themeKey:"palette",transform:vs});jl(lD,dD,uD);function kt(e){return e<=1&&e!==0?`${e*100}%`:e}const cD=Fe({prop:"width",transform:kt}),Pp=e=>{if(e.maxWidth!==void 0&&e.maxWidth!==null){const t=a=>{var i,s;const r=((i=e.theme)==null||(i=i.breakpoints)==null||(i=i.values)==null?void 0:i[a])||xp[a];return r?((s=e.theme)==null||(s=s.breakpoints)==null?void 0:s.unit)!=="px"?{maxWidth:`${r}${e.theme.breakpoints.unit}`}:{maxWidth:r}:{maxWidth:kt(a)}};return Na(e,e.maxWidth,t)}return null};Pp.filterProps=["maxWidth"];const pD=Fe({prop:"minWidth",transform:kt}),fD=Fe({prop:"height",transform:kt}),mD=Fe({prop:"maxHeight",transform:kt}),hD=Fe({prop:"minHeight",transform:kt});Fe({prop:"size",cssProperty:"width",transform:kt});Fe({prop:"size",cssProperty:"height",transform:kt});const gD=Fe({prop:"boxSizing"});jl(cD,Pp,pD,fD,mD,hD,gD);const yD={border:{themeKey:"borders",transform:Wt},borderTop:{themeKey:"borders",transform:Wt},borderRight:{themeKey:"borders",transform:Wt},borderBottom:{themeKey:"borders",transform:Wt},borderLeft:{themeKey:"borders",transform:Wt},borderColor:{themeKey:"palette"},borderTopColor:{themeKey:"palette"},borderRightColor:{themeKey:"palette"},borderBottomColor:{themeKey:"palette"},borderLeftColor:{themeKey:"palette"},outline:{themeKey:"borders",transform:Wt},outlineColor:{themeKey:"palette"},borderRadius:{themeKey:"shape.borderRadius",style:El},color:{themeKey:"palette",transform:vs},bgcolor:{themeKey:"palette",cssProperty:"backgroundColor",transform:vs},backgroundColor:{themeKey:"palette",transform:vs},p:{style:Re},pt:{style:Re},pr:{style:Re},pb:{style:Re},pl:{style:Re},px:{style:Re},py:{style:Re},padding:{style:Re},paddingTop:{style:Re},paddingRight:{style:Re},paddingBottom:{style:Re},paddingLeft:{style:Re},paddingX:{style:Re},paddingY:{style:Re},paddingInline:{style:Re},paddingInlineStart:{style:Re},paddingInlineEnd:{style:Re},paddingBlock:{style:Re},paddingBlockStart:{style:Re},paddingBlockEnd:{style:Re},m:{style:Ae},mt:{style:Ae},mr:{style:Ae},mb:{style:Ae},ml:{style:Ae},mx:{style:Ae},my:{style:Ae},margin:{style:Ae},marginTop:{style:Ae},marginRight:{style:Ae},marginBottom:{style:Ae},marginLeft:{style:Ae},marginX:{style:Ae},marginY:{style:Ae},marginInline:{style:Ae},marginInlineStart:{style:Ae},marginInlineEnd:{style:Ae},marginBlock:{style:Ae},marginBlockStart:{style:Ae},marginBlockEnd:{style:Ae},displayPrint:{cssProperty:!1,transform:e=>({"@media print":{display:e}})},display:{},overflow:{},textOverflow:{},visibility:{},whiteSpace:{},flexBasis:{},flexDirection:{},flexWrap:{},justifyContent:{},alignItems:{},alignContent:{},order:{},flex:{},flexGrow:{},flexShrink:{},alignSelf:{},justifyItems:{},justifySelf:{},gap:{style:Fl},rowGap:{style:Ol},columnGap:{style:Bl},gridColumn:{},gridRow:{},gridAutoFlow:{},gridAutoColumns:{},gridAutoRows:{},gridTemplateColumns:{},gridTemplateRows:{},gridTemplateAreas:{},gridArea:{},position:{},zIndex:{themeKey:"zIndex"},top:{},right:{},bottom:{},left:{},boxShadow:{themeKey:"shadows"},width:{transform:kt},maxWidth:{style:Pp},minWidth:{transform:kt},height:{transform:kt},maxHeight:{transform:kt},minHeight:{transform:kt},boxSizing:{},fontFamily:{themeKey:"typography"},fontSize:{themeKey:"typography"},fontStyle:{themeKey:"typography"},fontWeight:{themeKey:"typography"},letterSpacing:{},textTransform:{},lineHeight:{},textAlign:{},typography:{cssProperty:!1,themeKey:"typography"}},rn=yD;function _D(...e){const t=e.reduce((i,s)=>i.concat(Object.keys(s)),[]),a=new Set(t);return e.every(i=>a.size===Object.keys(i).length)}function vD(e,t){return typeof e=="function"?e(t):e}function yv(){function e(a,i,s,r){const n={[a]:i,theme:s},o=r[a];if(!o)return{[a]:i};const{cssProperty:l=a,themeKey:d,transform:c,style:p}=o;if(i==null)return null;if(d==="typography"&&i==="inherit")return{[a]:i};const u=Nl(s,d)||{};return p?p(n):Na(n,i,f=>{let h=Eo(u,c,f);return f===h&&typeof f=="string"&&(h=Eo(u,c,`${a}${f==="default"?"":X(f)}`,f)),l===!1?h:{[l]:h}})}function t(a){var i;const{sx:s,theme:r={}}=a||{};if(!s)return null;const n=(i=r.unstable_sxConfig)!=null?i:rn;function o(l){let d=l;if(typeof l=="function")d=l(r);else if(typeof l!="object")return l;if(!d)return null;const c=Iq(r.breakpoints),p=Object.keys(c);let u=c;return Object.keys(d).forEach(v=>{const f=vD(d[v],r);if(f!=null)if(typeof f=="object")if(n[v])u=vr(u,e(v,f,r,n));else{const h=Na({theme:r},f,T=>({[v]:T}));_D(h,f)?u[v]=t({sx:f,theme:r}):u=vr(u,h)}else u=vr(u,e(v,f,r,n))}),Nq(p,u)}return Array.isArray(s)?s.map(o):o(s)}return t}const _v=yv();_v.filterProps=["sx"];const Ml=_v;function vv(e,t){const a=this;return a.vars&&typeof a.getColorSchemeSelector=="function"?{[a.getColorSchemeSelector(e).replace(/(\[[^\]]+\])/,"*:where($1)")]:t}:a.palette.mode===e?t:{}}const bD=["breakpoints","palette","spacing","shape"];function Ap(e={},...t){const{breakpoints:a={},palette:i={},spacing:s,shape:r={}}=e,n=U(e,bD),o=mv(a),l=Lq(s);let d=xt({breakpoints:o,direction:"ltr",components:{},palette:q({mode:"light"},i),spacing:l,shape:q({},Rq,r)},n);return d.applyStyles=vv,d=t.reduce((c,p)=>xt(c,p),d),d.unstable_sxConfig=q({},rn,n==null?void 0:n.unstable_sxConfig),d.unstable_sx=function(p){return Ml({sx:p,theme:this})},d}const wD=Object.freeze(Object.defineProperty({__proto__:null,default:Ap,private_createBreakpoints:mv,unstable_applyStyles:vv},Symbol.toStringTag,{value:"Module"})),TD=Fa(wD),kD=["sx"],qD=e=>{var t,a;const i={systemProps:{},otherProps:{}},s=(t=e==null||(a=e.theme)==null?void 0:a.unstable_sxConfig)!=null?t:rn;return Object.keys(e).forEach(r=>{s[r]?i.systemProps[r]=e[r]:i.otherProps[r]=e[r]}),i};function Rp(e){const{sx:t}=e,a=U(e,kD),{systemProps:i,otherProps:s}=qD(a);let r;return Array.isArray(t)?r=[i,...t]:typeof t=="function"?r=(...n)=>{const o=t(...n);return Xa(o)?q({},i,o):i}:r=q({},i,t),q({},s,{sx:r})}const DD=Object.freeze(Object.defineProperty({__proto__:null,default:Ml,extendSxProp:Rp,unstable_createStyleFunctionSx:yv,unstable_defaultSxConfig:rn},Symbol.toStringTag,{value:"Module"})),xD=Fa(DD);var Bs=Zr;Object.defineProperty(en,"__esModule",{value:!0});var CD=en.default=LD;en.shouldForwardProp=to;en.systemDefaultTheme=void 0;var Bt=Bs(K_()),Qu=Bs(hk()),Jm=jD(qq),SD=Dq;Bs(xq);Bs(Cq);var PD=Bs(TD),AD=Bs(xD);const RD=["ownerState"],ID=["variants"],ND=["name","slot","skipVariantsResolver","skipSx","overridesResolver"];function bv(e){if(typeof WeakMap!="function")return null;var t=new WeakMap,a=new WeakMap;return(bv=function(i){return i?a:t})(e)}function jD(e,t){if(!t&&e&&e.__esModule)return e;if(e===null||typeof e!="object"&&typeof e!="function")return{default:e};var a=bv(t);if(a&&a.has(e))return a.get(e);var i={__proto__:null},s=Object.defineProperty&&Object.getOwnPropertyDescriptor;for(var r in e)if(r!=="default"&&Object.prototype.hasOwnProperty.call(e,r)){var n=s?Object.getOwnPropertyDescriptor(e,r):null;n&&(n.get||n.set)?Object.defineProperty(i,r,n):i[r]=e[r]}return i.default=e,a&&a.set(e,i),i}function ED(e){return Object.keys(e).length===0}function FD(e){return typeof e=="string"&&e.charCodeAt(0)>96}function to(e){return e!=="ownerState"&&e!=="theme"&&e!=="sx"&&e!=="as"}const BD=en.systemDefaultTheme=(0,PD.default)(),OD=e=>e&&e.charAt(0).toLowerCase()+e.slice(1);function An({defaultTheme:e,theme:t,themeId:a}){return ED(t)?e:t[a]||t}function MD(e){return e?(t,a)=>a[e]:null}function ao(e,t){let{ownerState:a}=t,i=(0,Qu.default)(t,RD);const s=typeof e=="function"?e((0,Bt.default)({ownerState:a},i)):e;if(Array.isArray(s))return s.flatMap(r=>ao(r,(0,Bt.default)({ownerState:a},i)));if(s&&typeof s=="object"&&Array.isArray(s.variants)){const{variants:r=[]}=s;let o=(0,Qu.default)(s,ID);return r.forEach(l=>{let d=!0;typeof l.props=="function"?d=l.props((0,Bt.default)({ownerState:a},i,a)):Object.keys(l.props).forEach(c=>{(a==null?void 0:a[c])!==l.props[c]&&i[c]!==l.props[c]&&(d=!1)}),d&&(Array.isArray(o)||(o=[o]),o.push(typeof l.style=="function"?l.style((0,Bt.default)({ownerState:a},i,a)):l.style))}),o}return s}function LD(e={}){const{themeId:t,defaultTheme:a=BD,rootShouldForwardProp:i=to,slotShouldForwardProp:s=to}=e,r=n=>(0,AD.default)((0,Bt.default)({},n,{theme:An((0,Bt.default)({},n,{defaultTheme:a,themeId:t}))}));return r.__mui_systemSx=!0,(n,o={})=>{(0,Jm.internal_processStyles)(n,D=>D.filter(x=>!(x!=null&&x.__mui_systemSx)));const{name:l,slot:d,skipVariantsResolver:c,skipSx:p,overridesResolver:u=MD(OD(d))}=o,v=(0,Qu.default)(o,ND),f=c!==void 0?c:d&&d!=="Root"&&d!=="root"||!1,h=p||!1;let T,y=to;d==="Root"||d==="root"?y=i:d?y=s:FD(n)&&(y=void 0);const g=(0,Jm.default)(n,(0,Bt.default)({shouldForwardProp:y,label:T},v)),_=D=>typeof D=="function"&&D.__emotion_real!==D||(0,SD.isPlainObject)(D)?x=>ao(D,(0,Bt.default)({},x,{theme:An({theme:x.theme,defaultTheme:a,themeId:t})})):D,w=(D,...x)=>{let k=_(D);const S=x?x.map(_):[];l&&u&&S.push(j=>{const E=An((0,Bt.default)({},j,{defaultTheme:a,themeId:t}));if(!E.components||!E.components[l]||!E.components[l].styleOverrides)return null;const F=E.components[l].styleOverrides,I={};return Object.entries(F).forEach(([B,O])=>{I[B]=ao(O,(0,Bt.default)({},j,{theme:E}))}),u(j,I)}),l&&!f&&S.push(j=>{var E;const F=An((0,Bt.default)({},j,{defaultTheme:a,themeId:t})),I=F==null||(E=F.components)==null||(E=E[l])==null?void 0:E.variants;return ao({variants:I},(0,Bt.default)({},j,{theme:F}))}),h||S.push(r);const A=S.length-x.length;if(Array.isArray(D)&&A>0){const j=new Array(A).fill("");k=[...D,...j],k.raw=[...D.raw,...j]}const P=g(k,...S);return n.muiName&&(P.muiName=n.muiName),P};return g.withConfig&&(w.withConfig=g.withConfig),w}}function WD(e,t){return q({toolbar:{minHeight:56,[e.up("xs")]:{"@media (orientation: landscape)":{minHeight:48}},[e.up("sm")]:{minHeight:64}}},t)}var Be={};const zD=Fa(OT),$D=Fa(uk);var wv=Zr;Object.defineProperty(Be,"__esModule",{value:!0});var xa=Be.alpha=Dv;Be.blend=ex;Be.colorChannel=void 0;var Yu=Be.darken=Np;Be.decomposeColor=Vt;Be.emphasize=xv;var UD=Be.getContrastRatio=XD;Be.getLuminance=Fo;Be.hexToRgb=Tv;Be.hslToRgb=qv;var Zu=Be.lighten=jp;Be.private_safeAlpha=JD;Be.private_safeColorChannel=void 0;Be.private_safeDarken=QD;Be.private_safeEmphasize=ZD;Be.private_safeLighten=YD;Be.recomposeColor=Os;Be.rgbToHex=GD;var Qm=wv(zD),KD=wv($D);function Ip(e,t=0,a=1){return(0,KD.default)(e,t,a)}function Tv(e){e=e.slice(1);const t=new RegExp(`.{1,${e.length>=6?2:1}}`,"g");let a=e.match(t);return a&&a[0].length===1&&(a=a.map(i=>i+i)),a?`rgb${a.length===4?"a":""}(${a.map((i,s)=>s<3?parseInt(i,16):Math.round(parseInt(i,16)/255*1e3)/1e3).join(", ")})`:""}function VD(e){const t=e.toString(16);return t.length===1?`0${t}`:t}function Vt(e){if(e.type)return e;if(e.charAt(0)==="#")return Vt(Tv(e));const t=e.indexOf("("),a=e.substring(0,t);if(["rgb","rgba","hsl","hsla","color"].indexOf(a)===-1)throw new Error((0,Qm.default)(9,e));let i=e.substring(t+1,e.length-1),s;if(a==="color"){if(i=i.split(" "),s=i.shift(),i.length===4&&i[3].charAt(0)==="/"&&(i[3]=i[3].slice(1)),["srgb","display-p3","a98-rgb","prophoto-rgb","rec-2020"].indexOf(s)===-1)throw new Error((0,Qm.default)(10,s))}else i=i.split(",");return i=i.map(r=>parseFloat(r)),{type:a,values:i,colorSpace:s}}const kv=e=>{const t=Vt(e);return t.values.slice(0,3).map((a,i)=>t.type.indexOf("hsl")!==-1&&i!==0?`${a}%`:a).join(" ")};Be.colorChannel=kv;const HD=(e,t)=>{try{return kv(e)}catch{return e}};Be.private_safeColorChannel=HD;function Os(e){const{type:t,colorSpace:a}=e;let{values:i}=e;return t.indexOf("rgb")!==-1?i=i.map((s,r)=>r<3?parseInt(s,10):s):t.indexOf("hsl")!==-1&&(i[1]=`${i[1]}%`,i[2]=`${i[2]}%`),t.indexOf("color")!==-1?i=`${a} ${i.join(" ")}`:i=`${i.join(", ")}`,`${t}(${i})`}function GD(e){if(e.indexOf("#")===0)return e;const{values:t}=Vt(e);return`#${t.map((a,i)=>VD(i===3?Math.round(255*a):a)).join("")}`}function qv(e){e=Vt(e);const{values:t}=e,a=t[0],i=t[1]/100,s=t[2]/100,r=i*Math.min(s,1-s),n=(d,c=(d+a/30)%12)=>s-r*Math.max(Math.min(c-3,9-c,1),-1);let o="rgb";const l=[Math.round(n(0)*255),Math.round(n(8)*255),Math.round(n(4)*255)];return e.type==="hsla"&&(o+="a",l.push(t[3])),Os({type:o,values:l})}function Fo(e){e=Vt(e);let t=e.type==="hsl"||e.type==="hsla"?Vt(qv(e)).values:e.values;return t=t.map(a=>(e.type!=="color"&&(a/=255),a<=.03928?a/12.92:((a+.055)/1.055)**2.4)),Number((.2126*t[0]+.7152*t[1]+.0722*t[2]).toFixed(3))}function XD(e,t){const a=Fo(e),i=Fo(t);return(Math.max(a,i)+.05)/(Math.min(a,i)+.05)}function Dv(e,t){return e=Vt(e),t=Ip(t),(e.type==="rgb"||e.type==="hsl")&&(e.type+="a"),e.type==="color"?e.values[3]=`/${t}`:e.values[3]=t,Os(e)}function JD(e,t,a){try{return Dv(e,t)}catch{return e}}function Np(e,t){if(e=Vt(e),t=Ip(t),e.type.indexOf("hsl")!==-1)e.values[2]*=1-t;else if(e.type.indexOf("rgb")!==-1||e.type.indexOf("color")!==-1)for(let a=0;a<3;a+=1)e.values[a]*=1-t;return Os(e)}function QD(e,t,a){try{return Np(e,t)}catch{return e}}function jp(e,t){if(e=Vt(e),t=Ip(t),e.type.indexOf("hsl")!==-1)e.values[2]+=(100-e.values[2])*t;else if(e.type.indexOf("rgb")!==-1)for(let a=0;a<3;a+=1)e.values[a]+=(255-e.values[a])*t;else if(e.type.indexOf("color")!==-1)for(let a=0;a<3;a+=1)e.values[a]+=(1-e.values[a])*t;return Os(e)}function YD(e,t,a){try{return jp(e,t)}catch{return e}}function xv(e,t=.15){return Fo(e)>.5?Np(e,t):jp(e,t)}function ZD(e,t,a){try{return xv(e,t)}catch{return e}}function ex(e,t,a,i=1){const s=(l,d)=>Math.round((l**(1/i)*(1-a)+d**(1/i)*a)**i),r=Vt(e),n=Vt(t),o=[s(r.values[0],n.values[0]),s(r.values[1],n.values[1]),s(r.values[2],n.values[2])];return Os({type:"rgb",values:o})}const tx={black:"#000",white:"#fff"},Kr=tx,ax={50:"#fafafa",100:"#f5f5f5",200:"#eeeeee",300:"#e0e0e0",400:"#bdbdbd",500:"#9e9e9e",600:"#757575",700:"#616161",800:"#424242",900:"#212121",A100:"#f5f5f5",A200:"#eeeeee",A400:"#bdbdbd",A700:"#616161"},ix=ax,sx={50:"#f3e5f5",100:"#e1bee7",200:"#ce93d8",300:"#ba68c8",400:"#ab47bc",500:"#9c27b0",600:"#8e24aa",700:"#7b1fa2",800:"#6a1b9a",900:"#4a148c",A100:"#ea80fc",A200:"#e040fb",A400:"#d500f9",A700:"#aa00ff"},Vi=sx,rx={50:"#ffebee",100:"#ffcdd2",200:"#ef9a9a",300:"#e57373",400:"#ef5350",500:"#f44336",600:"#e53935",700:"#d32f2f",800:"#c62828",900:"#b71c1c",A100:"#ff8a80",A200:"#ff5252",A400:"#ff1744",A700:"#d50000"},Hi=rx,nx={50:"#fff3e0",100:"#ffe0b2",200:"#ffcc80",300:"#ffb74d",400:"#ffa726",500:"#ff9800",600:"#fb8c00",700:"#f57c00",800:"#ef6c00",900:"#e65100",A100:"#ffd180",A200:"#ffab40",A400:"#ff9100",A700:"#ff6d00"},Qs=nx,ox={50:"#e3f2fd",100:"#bbdefb",200:"#90caf9",300:"#64b5f6",400:"#42a5f5",500:"#2196f3",600:"#1e88e5",700:"#1976d2",800:"#1565c0",900:"#0d47a1",A100:"#82b1ff",A200:"#448aff",A400:"#2979ff",A700:"#2962ff"},Gi=ox,lx={50:"#e1f5fe",100:"#b3e5fc",200:"#81d4fa",300:"#4fc3f7",400:"#29b6f6",500:"#03a9f4",600:"#039be5",700:"#0288d1",800:"#0277bd",900:"#01579b",A100:"#80d8ff",A200:"#40c4ff",A400:"#00b0ff",A700:"#0091ea"},Xi=lx,dx={50:"#e8f5e9",100:"#c8e6c9",200:"#a5d6a7",300:"#81c784",400:"#66bb6a",500:"#4caf50",600:"#43a047",700:"#388e3c",800:"#2e7d32",900:"#1b5e20",A100:"#b9f6ca",A200:"#69f0ae",A400:"#00e676",A700:"#00c853"},Ji=dx,ux=["mode","contrastThreshold","tonalOffset"],Ym={text:{primary:"rgba(0, 0, 0, 0.87)",secondary:"rgba(0, 0, 0, 0.6)",disabled:"rgba(0, 0, 0, 0.38)"},divider:"rgba(0, 0, 0, 0.12)",background:{paper:Kr.white,default:Kr.white},action:{active:"rgba(0, 0, 0, 0.54)",hover:"rgba(0, 0, 0, 0.04)",hoverOpacity:.04,selected:"rgba(0, 0, 0, 0.08)",selectedOpacity:.08,disabled:"rgba(0, 0, 0, 0.26)",disabledBackground:"rgba(0, 0, 0, 0.12)",disabledOpacity:.38,focus:"rgba(0, 0, 0, 0.12)",focusOpacity:.12,activatedOpacity:.12}},Id={text:{primary:Kr.white,secondary:"rgba(255, 255, 255, 0.7)",disabled:"rgba(255, 255, 255, 0.5)",icon:"rgba(255, 255, 255, 0.5)"},divider:"rgba(255, 255, 255, 0.12)",background:{paper:"#121212",default:"#121212"},action:{active:Kr.white,hover:"rgba(255, 255, 255, 0.08)",hoverOpacity:.08,selected:"rgba(255, 255, 255, 0.16)",selectedOpacity:.16,disabled:"rgba(255, 255, 255, 0.3)",disabledBackground:"rgba(255, 255, 255, 0.12)",disabledOpacity:.38,focus:"rgba(255, 255, 255, 0.12)",focusOpacity:.12,activatedOpacity:.24}};function Zm(e,t,a,i){const s=i.light||i,r=i.dark||i*1.5;e[t]||(e.hasOwnProperty(a)?e[t]=e[a]:t==="light"?e.light=Zu(e.main,s):t==="dark"&&(e.dark=Yu(e.main,r)))}function cx(e="light"){return e==="dark"?{main:Gi[200],light:Gi[50],dark:Gi[400]}:{main:Gi[700],light:Gi[400],dark:Gi[800]}}function px(e="light"){return e==="dark"?{main:Vi[200],light:Vi[50],dark:Vi[400]}:{main:Vi[500],light:Vi[300],dark:Vi[700]}}function fx(e="light"){return e==="dark"?{main:Hi[500],light:Hi[300],dark:Hi[700]}:{main:Hi[700],light:Hi[400],dark:Hi[800]}}function mx(e="light"){return e==="dark"?{main:Xi[400],light:Xi[300],dark:Xi[700]}:{main:Xi[700],light:Xi[500],dark:Xi[900]}}function hx(e="light"){return e==="dark"?{main:Ji[400],light:Ji[300],dark:Ji[700]}:{main:Ji[800],light:Ji[500],dark:Ji[900]}}function gx(e="light"){return e==="dark"?{main:Qs[400],light:Qs[300],dark:Qs[700]}:{main:"#ed6c02",light:Qs[500],dark:Qs[900]}}function yx(e){const{mode:t="light",contrastThreshold:a=3,tonalOffset:i=.2}=e,s=U(e,ux),r=e.primary||cx(t),n=e.secondary||px(t),o=e.error||fx(t),l=e.info||mx(t),d=e.success||hx(t),c=e.warning||gx(t);function p(h){return UD(h,Id.text.primary)>=a?Id.text.primary:Ym.text.primary}const u=({color:h,name:T,mainShade:y=500,lightShade:g=300,darkShade:_=700})=>{if(h=q({},h),!h.main&&h[y]&&(h.main=h[y]),!h.hasOwnProperty("main"))throw new Error(Oi(11,T?` (${T})`:"",y));if(typeof h.main!="string")throw new Error(Oi(12,T?` (${T})`:"",JSON.stringify(h.main)));return Zm(h,"light",g,i),Zm(h,"dark",_,i),h.contrastText||(h.contrastText=p(h.main)),h},v={dark:Id,light:Ym};return xt(q({common:q({},Kr),mode:t,primary:u({color:r,name:"primary"}),secondary:u({color:n,name:"secondary",mainShade:"A400",lightShade:"A200",darkShade:"A700"}),error:u({color:o,name:"error"}),warning:u({color:c,name:"warning"}),info:u({color:l,name:"info"}),success:u({color:d,name:"success"}),grey:ix,contrastThreshold:a,getContrastText:p,augmentColor:u,tonalOffset:i},v[t]),s)}const _x=["fontFamily","fontSize","fontWeightLight","fontWeightRegular","fontWeightMedium","fontWeightBold","htmlFontSize","allVariants","pxToRem"];function vx(e){return Math.round(e*1e5)/1e5}const eh={textTransform:"uppercase"},th='"Roboto", "Helvetica", "Arial", sans-serif';function bx(e,t){const a=typeof t=="function"?t(e):t,{fontFamily:i=th,fontSize:s=14,fontWeightLight:r=300,fontWeightRegular:n=400,fontWeightMedium:o=500,fontWeightBold:l=700,htmlFontSize:d=16,allVariants:c,pxToRem:p}=a,u=U(a,_x),v=s/14,f=p||(y=>`${y/d*v}rem`),h=(y,g,_,w,D)=>q({fontFamily:i,fontWeight:y,fontSize:f(g),lineHeight:_},i===th?{letterSpacing:`${vx(w/g)}em`}:{},D,c),T={h1:h(r,96,1.167,-1.5),h2:h(r,60,1.2,-.5),h3:h(n,48,1.167,0),h4:h(n,34,1.235,.25),h5:h(n,24,1.334,0),h6:h(o,20,1.6,.15),subtitle1:h(n,16,1.75,.15),subtitle2:h(o,14,1.57,.1),body1:h(n,16,1.5,.15),body2:h(n,14,1.43,.15),button:h(o,14,1.75,.4,eh),caption:h(n,12,1.66,.4),overline:h(n,12,2.66,1,eh),inherit:{fontFamily:"inherit",fontWeight:"inherit",fontSize:"inherit",lineHeight:"inherit",letterSpacing:"inherit"}};return xt(q({htmlFontSize:d,pxToRem:f,fontFamily:i,fontSize:s,fontWeightLight:r,fontWeightRegular:n,fontWeightMedium:o,fontWeightBold:l},T),u,{clone:!1})}const wx=.2,Tx=.14,kx=.12;function ke(...e){return[`${e[0]}px ${e[1]}px ${e[2]}px ${e[3]}px rgba(0,0,0,${wx})`,`${e[4]}px ${e[5]}px ${e[6]}px ${e[7]}px rgba(0,0,0,${Tx})`,`${e[8]}px ${e[9]}px ${e[10]}px ${e[11]}px rgba(0,0,0,${kx})`].join(",")}const qx=["none",ke(0,2,1,-1,0,1,1,0,0,1,3,0),ke(0,3,1,-2,0,2,2,0,0,1,5,0),ke(0,3,3,-2,0,3,4,0,0,1,8,0),ke(0,2,4,-1,0,4,5,0,0,1,10,0),ke(0,3,5,-1,0,5,8,0,0,1,14,0),ke(0,3,5,-1,0,6,10,0,0,1,18,0),ke(0,4,5,-2,0,7,10,1,0,2,16,1),ke(0,5,5,-3,0,8,10,1,0,3,14,2),ke(0,5,6,-3,0,9,12,1,0,3,16,2),ke(0,6,6,-3,0,10,14,1,0,4,18,3),ke(0,6,7,-4,0,11,15,1,0,4,20,3),ke(0,7,8,-4,0,12,17,2,0,5,22,4),ke(0,7,8,-4,0,13,19,2,0,5,24,4),ke(0,7,9,-4,0,14,21,2,0,5,26,4),ke(0,8,9,-5,0,15,22,2,0,6,28,5),ke(0,8,10,-5,0,16,24,2,0,6,30,5),ke(0,8,11,-5,0,17,26,2,0,6,32,5),ke(0,9,11,-5,0,18,28,2,0,7,34,6),ke(0,9,12,-6,0,19,29,2,0,7,36,6),ke(0,10,13,-6,0,20,31,3,0,8,38,7),ke(0,10,13,-6,0,21,33,3,0,8,40,7),ke(0,10,14,-6,0,22,35,3,0,8,42,7),ke(0,11,14,-7,0,23,36,3,0,9,44,8),ke(0,11,15,-7,0,24,38,3,0,9,46,8)],Dx=qx,xx=["duration","easing","delay"],Cx={easeInOut:"cubic-bezier(0.4, 0, 0.2, 1)",easeOut:"cubic-bezier(0.0, 0, 0.2, 1)",easeIn:"cubic-bezier(0.4, 0, 1, 1)",sharp:"cubic-bezier(0.4, 0, 0.6, 1)"},Sx={shortest:150,shorter:200,short:250,standard:300,complex:375,enteringScreen:225,leavingScreen:195};function ah(e){return`${Math.round(e)}ms`}function Px(e){if(!e)return 0;const t=e/36;return Math.round((4+15*t**.25+t/5)*10)}function Ax(e){const t=q({},Cx,e.easing),a=q({},Sx,e.duration);return q({getAutoHeightDuration:Px,create:(s=["all"],r={})=>{const{duration:n=a.standard,easing:o=t.easeInOut,delay:l=0}=r;return U(r,xx),(Array.isArray(s)?s:[s]).map(d=>`${d} ${typeof n=="string"?n:ah(n)} ${o} ${typeof l=="string"?l:ah(l)}`).join(",")}},e,{easing:t,duration:a})}const Rx={mobileStepper:1e3,fab:1050,speedDial:1050,appBar:1100,drawer:1200,modal:1300,snackbar:1400,tooltip:1500},Ix=Rx,Nx=["breakpoints","mixins","spacing","palette","transitions","typography","shape"];function Cv(e={},...t){const{mixins:a={},palette:i={},transitions:s={},typography:r={}}=e,n=U(e,Nx);if(e.vars)throw new Error(Oi(18));const o=yx(i),l=Ap(e);let d=xt(l,{mixins:WD(l.breakpoints,a),palette:o,shadows:Dx.slice(),typography:bx(o,r),transitions:Ax(s),zIndex:q({},Ix)});return d=xt(d,n),d=t.reduce((c,p)=>xt(c,p),d),d.unstable_sxConfig=q({},rn,n==null?void 0:n.unstable_sxConfig),d.unstable_sx=function(p){return Ml({sx:p,theme:this})},d}const jx=Cv(),Ep=jx,Ll="$$material";function Sv(e){return e!=="ownerState"&&e!=="theme"&&e!=="sx"&&e!=="as"}const Ex=e=>Sv(e)&&e!=="classes",It=Ex,Fx=CD({themeId:Ll,defaultTheme:Ep,rootShouldForwardProp:It}),z=Fx;function Bx(e){return pe("MuiSvgIcon",e)}re("MuiSvgIcon",["root","colorPrimary","colorSecondary","colorAction","colorError","colorDisabled","fontSizeInherit","fontSizeSmall","fontSizeMedium","fontSizeLarge"]);const Ox=["children","className","color","component","fontSize","htmlColor","inheritViewBox","titleAccess","viewBox"],Mx=e=>{const{color:t,fontSize:a,classes:i}=e,s={root:["root",t!=="inherit"&&`color${X(t)}`,`fontSize${X(a)}`]};return fe(s,Bx,i)},Lx=z("svg",{name:"MuiSvgIcon",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,a.color!=="inherit"&&t[`color${X(a.color)}`],t[`fontSize${X(a.fontSize)}`]]}})(({theme:e,ownerState:t})=>{var a,i,s,r,n,o,l,d,c,p,u,v,f;return{userSelect:"none",width:"1em",height:"1em",display:"inline-block",fill:t.hasSvgAsChild?void 0:"currentColor",flexShrink:0,transition:(a=e.transitions)==null||(i=a.create)==null?void 0:i.call(a,"fill",{duration:(s=e.transitions)==null||(s=s.duration)==null?void 0:s.shorter}),fontSize:{inherit:"inherit",small:((r=e.typography)==null||(n=r.pxToRem)==null?void 0:n.call(r,20))||"1.25rem",medium:((o=e.typography)==null||(l=o.pxToRem)==null?void 0:l.call(o,24))||"1.5rem",large:((d=e.typography)==null||(c=d.pxToRem)==null?void 0:c.call(d,35))||"2.1875rem"}[t.fontSize],color:(p=(u=(e.vars||e).palette)==null||(u=u[t.color])==null?void 0:u.main)!=null?p:{action:(v=(e.vars||e).palette)==null||(v=v.action)==null?void 0:v.active,disabled:(f=(e.vars||e).palette)==null||(f=f.action)==null?void 0:f.disabled,inherit:void 0}[t.color]}}),Pv=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiSvgIcon"}),{children:s,className:r,color:n="inherit",component:o="svg",fontSize:l="medium",htmlColor:d,inheritViewBox:c=!1,titleAccess:p,viewBox:u="0 0 24 24"}=i,v=U(i,Ox),f=b.isValidElement(s)&&s.type==="svg",h=q({},i,{color:n,component:o,fontSize:l,instanceFontSize:t.fontSize,inheritViewBox:c,viewBox:u,hasSvgAsChild:f}),T={};c||(T.viewBox=u);const y=Mx(h);return m.jsxs(Lx,q({as:o,className:H(y.root,r),focusable:"false",color:d,"aria-hidden":p?void 0:!0,role:p?"img":void 0,ref:a},T,v,f&&s.props,{ownerState:h,children:[f?s.props.children:s,p?m.jsx("title",{children:p}):null]}))});Pv.muiName="SvgIcon";const ih=Pv;function zi(e,t){function a(i,s){return m.jsx(ih,q({"data-testid":`${t}Icon`,ref:s},i,{children:e}))}return a.muiName=ih.muiName,b.memo(b.forwardRef(a))}const Wx={configure:e=>{_p.configure(e)}},zx=Object.freeze(Object.defineProperty({__proto__:null,capitalize:X,createChainedFunction:$u,createSvgIcon:zi,debounce:hl,deprecatedPropType:KT,isMuiElement:Qn,ownerDocument:_t,ownerWindow:ha,requirePropFactory:VT,setRef:Io,unstable_ClassNameGenerator:Wx,unstable_useEnhancedEffect:Ia,unstable_useId:yp,unsupportedProp:XT,useControlled:Uu,useEventCallback:Si,useForkRef:Ke,useIsFocusVisible:O_},Symbol.toStringTag,{value:"Module"})),$x=Fa(zx);var sh;function Fp(){return sh||(sh=1,function(e){"use client";Object.defineProperty(e,"__esModule",{value:!0}),Object.defineProperty(e,"default",{enumerable:!0,get:function(){return t.createSvgIcon}});var t=$x}(Cd)),Cd}var Ux=Zr;Object.defineProperty(mp,"__esModule",{value:!0});var Av=mp.default=void 0,Kx=Ux(Fp()),Vx=m;Av=mp.default=(0,Kx.default)((0,Vx.jsx)("path",{d:"M15.5 14h-.79l-.28-.27C15.41 12.59 16 11.11 16 9.5 16 5.91 13.09 3 9.5 3S3 5.91 3 9.5 5.91 16 9.5 16c1.61 0 3.09-.59 4.23-1.57l.27.28v.79l5 4.99L20.49 19zm-6 0C7.01 14 5 11.99 5 9.5S7.01 5 9.5 5 14 7.01 14 9.5 11.99 14 9.5 14"}),"Search");var Bp={},Hx=Zr;Object.defineProperty(Bp,"__esModule",{value:!0});var Rv=Bp.default=void 0,Gx=Hx(Fp()),Xx=m;Rv=Bp.default=(0,Gx.default)((0,Xx.jsx)("path",{d:"M19 7v4H5.83l3.58-3.59L8 6l-6 6 6 6 1.41-1.41L5.83 13H21V7z"}),"KeyboardReturn");function ja(e){return Array.isArray?Array.isArray(e):jv(e)==="[object Array]"}const Jx=1/0;function Qx(e){if(typeof e=="string")return e;let t=e+"";return t=="0"&&1/e==-Jx?"-0":t}function Yx(e){return e==null?"":Qx(e)}function ca(e){return typeof e=="string"}function Iv(e){return typeof e=="number"}function Zx(e){return e===!0||e===!1||eC(e)&&jv(e)=="[object Boolean]"}function Nv(e){return typeof e=="object"}function eC(e){return Nv(e)&&e!==null}function Tt(e){return e!=null}function Nd(e){return!e.trim().length}function jv(e){return e==null?e===void 0?"[object Undefined]":"[object Null]":Object.prototype.toString.call(e)}const tC="Incorrect 'index' type",aC=e=>`Invalid value for key ${e}`,iC=e=>`Pattern length exceeds max of ${e}.`,sC=e=>`Missing ${e} property in key`,rC=e=>`Property 'weight' in key '${e}' must be a positive integer`,rh=Object.prototype.hasOwnProperty;class nC{constructor(t){this._keys=[],this._keyMap={};let a=0;t.forEach(i=>{let s=Ev(i);a+=s.weight,this._keys.push(s),this._keyMap[s.id]=s,a+=s.weight}),this._keys.forEach(i=>{i.weight/=a})}get(t){return this._keyMap[t]}keys(){return this._keys}toJSON(){return JSON.stringify(this._keys)}}function Ev(e){let t=null,a=null,i=null,s=1,r=null;if(ca(e)||ja(e))i=e,t=nh(e),a=ec(e);else{if(!rh.call(e,"name"))throw new Error(sC("name"));const n=e.name;if(i=n,rh.call(e,"weight")&&(s=e.weight,s<=0))throw new Error(rC(n));t=nh(n),a=ec(n),r=e.getFn}return{path:t,id:a,weight:s,src:i,getFn:r}}function nh(e){return ja(e)?e:e.split(".")}function ec(e){return ja(e)?e.join("."):e}function oC(e,t){let a=[],i=!1;const s=(r,n,o)=>{if(Tt(r))if(!n[o])a.push(r);else{let l=n[o];const d=r[l];if(!Tt(d))return;if(o===n.length-1&&(ca(d)||Iv(d)||Zx(d)))a.push(Yx(d));else if(ja(d)){i=!0;for(let c=0,p=d.length;ce.score===t.score?e.idx{this._keysMap[a.id]=i})}create(){this.isCreated||!this.docs.length||(this.isCreated=!0,ca(this.docs[0])?this.docs.forEach((t,a)=>{this._addString(t,a)}):this.docs.forEach((t,a)=>{this._addObject(t,a)}),this.norm.clear())}add(t){const a=this.size();ca(t)?this._addString(t,a):this._addObject(t,a)}removeAt(t){this.records.splice(t,1);for(let a=t,i=this.size();a{let n=s.getFn?s.getFn(t):this.getFn(t,s.path);if(Tt(n)){if(ja(n)){let o=[];const l=[{nestedArrIndex:-1,value:n}];for(;l.length;){const{nestedArrIndex:d,value:c}=l.pop();if(Tt(c))if(ca(c)&&!Nd(c)){let p={v:c,i:d,n:this.norm.get(c)};o.push(p)}else ja(c)&&c.forEach((p,u)=>{l.push({nestedArrIndex:u,value:p})})}i.$[r]=o}else if(ca(n)&&!Nd(n)){let o={v:n,n:this.norm.get(n)};i.$[r]=o}}}),this.records.push(i)}toJSON(){return{keys:this.keys,records:this.records}}}function Fv(e,t,{getFn:a=G.getFn,fieldNormWeight:i=G.fieldNormWeight}={}){const s=new Op({getFn:a,fieldNormWeight:i});return s.setKeys(e.map(Ev)),s.setSources(t),s.create(),s}function mC(e,{getFn:t=G.getFn,fieldNormWeight:a=G.fieldNormWeight}={}){const{keys:i,records:s}=e,r=new Op({getFn:t,fieldNormWeight:a});return r.setKeys(i),r.setIndexRecords(s),r}function Rn(e,{errors:t=0,currentLocation:a=0,expectedLocation:i=0,distance:s=G.distance,ignoreLocation:r=G.ignoreLocation}={}){const n=t/e.length;if(r)return n;const o=Math.abs(i-a);return s?n+o/s:o?1:n}function hC(e=[],t=G.minMatchCharLength){let a=[],i=-1,s=-1,r=0;for(let n=e.length;r=t&&a.push([i,s]),i=-1)}return e[r-1]&&r-i>=t&&a.push([i,r-1]),a}const Ti=32;function gC(e,t,a,{location:i=G.location,distance:s=G.distance,threshold:r=G.threshold,findAllMatches:n=G.findAllMatches,minMatchCharLength:o=G.minMatchCharLength,includeMatches:l=G.includeMatches,ignoreLocation:d=G.ignoreLocation}={}){if(t.length>Ti)throw new Error(iC(Ti));const c=t.length,p=e.length,u=Math.max(0,Math.min(i,p));let v=r,f=u;const h=o>1||l,T=h?Array(p):[];let y;for(;(y=e.indexOf(t,f))>-1;){let k=Rn(t,{currentLocation:y,expectedLocation:u,distance:s,ignoreLocation:d});if(v=Math.min(k,v),f=y+c,h){let S=0;for(;S=P;I-=1){let B=I-1,O=a[e.charAt(B)];if(h&&(T[B]=+!!O),E[I]=(E[I+1]<<1|1)&O,k&&(E[I]|=(g[I+1]|g[I])<<1|1|g[I+1]),E[I]&D&&(_=Rn(t,{errors:k,currentLocation:B,expectedLocation:u,distance:s,ignoreLocation:d}),_<=v)){if(v=_,f=B,f<=u)break;P=Math.max(1,2*u-f)}}if(Rn(t,{errors:k+1,currentLocation:u,expectedLocation:u,distance:s,ignoreLocation:d})>v)break;g=E}const x={isMatch:f>=0,score:Math.max(.001,_)};if(h){const k=hC(T,o);k.length?l&&(x.indices=k):x.isMatch=!1}return x}function yC(e){let t={};for(let a=0,i=e.length;a{this.chunks.push({pattern:u,alphabet:yC(u),startIndex:v})},p=this.pattern.length;if(p>Ti){let u=0;const v=p%Ti,f=p-v;for(;u{const{isMatch:y,score:g,indices:_}=gC(t,f,h,{location:s+T,distance:r,threshold:n,findAllMatches:o,minMatchCharLength:l,includeMatches:i,ignoreLocation:d});y&&(u=!0),p+=g,y&&_&&(c=[...c,..._])});let v={isMatch:u,score:u?p/this.chunks.length:1};return u&&i&&(v.indices=c),v}}class mi{constructor(t){this.pattern=t}static isMultiMatch(t){return oh(t,this.multiRegex)}static isSingleMatch(t){return oh(t,this.singleRegex)}search(){}}function oh(e,t){const a=e.match(t);return a?a[1]:null}class _C extends mi{constructor(t){super(t)}static get type(){return"exact"}static get multiRegex(){return/^="(.*)"$/}static get singleRegex(){return/^=(.*)$/}search(t){const a=t===this.pattern;return{isMatch:a,score:a?0:1,indices:[0,this.pattern.length-1]}}}class vC extends mi{constructor(t){super(t)}static get type(){return"inverse-exact"}static get multiRegex(){return/^!"(.*)"$/}static get singleRegex(){return/^!(.*)$/}search(t){const i=t.indexOf(this.pattern)===-1;return{isMatch:i,score:i?0:1,indices:[0,t.length-1]}}}class bC extends mi{constructor(t){super(t)}static get type(){return"prefix-exact"}static get multiRegex(){return/^\^"(.*)"$/}static get singleRegex(){return/^\^(.*)$/}search(t){const a=t.startsWith(this.pattern);return{isMatch:a,score:a?0:1,indices:[0,this.pattern.length-1]}}}class wC extends mi{constructor(t){super(t)}static get type(){return"inverse-prefix-exact"}static get multiRegex(){return/^!\^"(.*)"$/}static get singleRegex(){return/^!\^(.*)$/}search(t){const a=!t.startsWith(this.pattern);return{isMatch:a,score:a?0:1,indices:[0,t.length-1]}}}class TC extends mi{constructor(t){super(t)}static get type(){return"suffix-exact"}static get multiRegex(){return/^"(.*)"\$$/}static get singleRegex(){return/^(.*)\$$/}search(t){const a=t.endsWith(this.pattern);return{isMatch:a,score:a?0:1,indices:[t.length-this.pattern.length,t.length-1]}}}class kC extends mi{constructor(t){super(t)}static get type(){return"inverse-suffix-exact"}static get multiRegex(){return/^!"(.*)"\$$/}static get singleRegex(){return/^!(.*)\$$/}search(t){const a=!t.endsWith(this.pattern);return{isMatch:a,score:a?0:1,indices:[0,t.length-1]}}}class Ov extends mi{constructor(t,{location:a=G.location,threshold:i=G.threshold,distance:s=G.distance,includeMatches:r=G.includeMatches,findAllMatches:n=G.findAllMatches,minMatchCharLength:o=G.minMatchCharLength,isCaseSensitive:l=G.isCaseSensitive,ignoreLocation:d=G.ignoreLocation}={}){super(t),this._bitapSearch=new Bv(t,{location:a,threshold:i,distance:s,includeMatches:r,findAllMatches:n,minMatchCharLength:o,isCaseSensitive:l,ignoreLocation:d})}static get type(){return"fuzzy"}static get multiRegex(){return/^"(.*)"$/}static get singleRegex(){return/^(.*)$/}search(t){return this._bitapSearch.searchIn(t)}}class Mv extends mi{constructor(t){super(t)}static get type(){return"include"}static get multiRegex(){return/^'"(.*)"$/}static get singleRegex(){return/^'(.*)$/}search(t){let a=0,i;const s=[],r=this.pattern.length;for(;(i=t.indexOf(this.pattern,a))>-1;)a=i+r,s.push([i,a-1]);const n=!!s.length;return{isMatch:n,score:n?0:1,indices:s}}}const tc=[_C,Mv,bC,wC,kC,TC,vC,Ov],lh=tc.length,qC=/ +(?=(?:[^\"]*\"[^\"]*\")*[^\"]*$)/,DC="|";function xC(e,t={}){return e.split(DC).map(a=>{let i=a.trim().split(qC).filter(r=>r&&!!r.trim()),s=[];for(let r=0,n=i.length;r!!(e[Bo.AND]||e[Bo.OR]),AC=e=>!!e[sc.PATH],RC=e=>!ja(e)&&Nv(e)&&!rc(e),dh=e=>({[Bo.AND]:Object.keys(e).map(t=>({[t]:e[t]}))});function Lv(e,t,{auto:a=!0}={}){const i=s=>{let r=Object.keys(s);const n=AC(s);if(!n&&r.length>1&&!rc(s))return i(dh(s));if(RC(s)){const l=n?s[sc.PATH]:r[0],d=n?s[sc.PATTERN]:s[l];if(!ca(d))throw new Error(aC(l));const c={keyId:ec(l),pattern:d};return a&&(c.searcher=ic(d,t)),c}let o={children:[],operator:r[0]};return r.forEach(l=>{const d=s[l];ja(d)&&d.forEach(c=>{o.children.push(i(c))})}),o};return rc(e)||(e=dh(e)),i(e)}function IC(e,{ignoreFieldNorm:t=G.ignoreFieldNorm}){e.forEach(a=>{let i=1;a.matches.forEach(({key:s,norm:r,score:n})=>{const o=s?s.weight:null;i*=Math.pow(n===0&&o?Number.EPSILON:n,(o||1)*(t?1:r))}),a.score=i})}function NC(e,t){const a=e.matches;t.matches=[],Tt(a)&&a.forEach(i=>{if(!Tt(i.indices)||!i.indices.length)return;const{indices:s,value:r}=i;let n={indices:s,value:r};i.key&&(n.key=i.key.src),i.idx>-1&&(n.refIndex=i.idx),t.matches.push(n)})}function jC(e,t){t.score=e.score}function EC(e,t,{includeMatches:a=G.includeMatches,includeScore:i=G.includeScore}={}){const s=[];return a&&s.push(NC),i&&s.push(jC),e.map(r=>{const{idx:n}=r,o={item:t[n],refIndex:n};return s.length&&s.forEach(l=>{l(r,o)}),o})}class Ms{constructor(t,a={},i){this.options={...G,...a},this.options.useExtendedSearch,this._keyStore=new nC(this.options.keys),this.setCollection(t,i)}setCollection(t,a){if(this._docs=t,a&&!(a instanceof Op))throw new Error(tC);this._myIndex=a||Fv(this.options.keys,this._docs,{getFn:this.options.getFn,fieldNormWeight:this.options.fieldNormWeight})}add(t){Tt(t)&&(this._docs.push(t),this._myIndex.add(t))}remove(t=()=>!1){const a=[];for(let i=0,s=this._docs.length;i-1&&(l=l.slice(0,a)),EC(l,this._docs,{includeMatches:i,includeScore:s})}_searchStringList(t){const a=ic(t,this.options),{records:i}=this._myIndex,s=[];return i.forEach(({v:r,i:n,n:o})=>{if(!Tt(r))return;const{isMatch:l,score:d,indices:c}=a.searchIn(r);l&&s.push({item:r,idx:n,matches:[{score:d,value:r,norm:o,indices:c}]})}),s}_searchLogical(t){const a=Lv(t,this.options),i=(o,l,d)=>{if(!o.children){const{keyId:p,searcher:u}=o,v=this._findMatches({key:this._keyStore.get(p),value:this._myIndex.getValueForItemAtKeyId(l,p),searcher:u});return v&&v.length?[{idx:d,item:l,matches:v}]:[]}const c=[];for(let p=0,u=o.children.length;p{if(Tt(o)){let d=i(a,o,l);d.length&&(r[l]||(r[l]={idx:l,item:o,matches:[]},n.push(r[l])),d.forEach(({matches:c})=>{r[l].matches.push(...c)}))}}),n}_searchObjectList(t){const a=ic(t,this.options),{keys:i,records:s}=this._myIndex,r=[];return s.forEach(({$:n,i:o})=>{if(!Tt(n))return;let l=[];i.forEach((d,c)=>{l.push(...this._findMatches({key:d,value:n[c],searcher:a}))}),l.length&&r.push({idx:o,item:n,matches:l})}),r}_findMatches({key:t,value:a,searcher:i}){if(!Tt(a))return[];let s=[];if(ja(a))a.forEach(({v:r,i:n,n:o})=>{if(!Tt(r))return;const{isMatch:l,score:d,indices:c}=i.searchIn(r);l&&s.push({score:d,key:t,value:r,idx:n,norm:o,indices:c})});else{const{v:r,n}=a,{isMatch:o,score:l,indices:d}=i.searchIn(r);o&&s.push({score:l,key:t,value:r,norm:n,indices:d})}return s}}Ms.version="6.6.2";Ms.createIndex=Fv;Ms.parseIndex=mC;Ms.config=G;Ms.parseQuery=Lv;PC(SC);const Wv=b.createContext(),nn=()=>b.useContext(Wv),FC=({children:e})=>{const[t,a]=b.useState(""),[i,s]=b.useState(),[r,n]=b.useState(!1);return m.jsx(Wv.Provider,{value:{searchQuery:t,setSearchQuery:a,searchResults:i,setSearchResults:s,isSearchSubmitted:r,setIsSearchSubmitted:n},children:e})};function zv(e,t){const a=JSON.parse(e);let i=[];function s(o,l){if(typeof l=="string"&&l.toLowerCase().includes(t.toLowerCase()))return l.trim();if(typeof l=="object")for(const d in l){const c=s(d,l[d]);d>0&&typeof c=="string"?i.push(o+": "+l[(d-"1").toString()]+" "+c):d==0&&l.length>1&&typeof c=="string"?i.push(o+": "+c+" "+l[1]):typeof c=="string"&&i.push(d+": "+c)}}for(const o in a)s(o,a[o]);const r=i[0];let n=[];try{const o=r.toLowerCase(),l=t.toLowerCase(),d=o.indexOf(l);if(d!==-1){const c=r.substring(0,d),p=r.substring(d,d+l.length),u=r.substring(d+l.length);n=[c,p,u]}}catch{n=[null,null,null]}return n}const BC=ya.plugins;let uh=["name","metadata.description","entry_point_prefix","metadata.author"];function OC(e){const t=[],a=JSON.parse(JSON.stringify(e));return Object.entries(a).forEach(([s,r])=>{Object.entries(r.entry_points).forEach(([n,o])=>{for(const l in o){let d=["entry_points",n,l];r.entry_points[n][l]=JSON.stringify(r.entry_points[n][l]),uh.push(d)}}),t.push(r)}),new Ms(t,{keys:uh,includeScore:!0,ignoreLocation:!0,threshold:.1,includeMatches:!0})}const MC=OC(BC);function LC(){const{searchQuery:e,setSearchQuery:t,setSearchResults:a,isSearchSubmitted:i,setIsSearchSubmitted:s}=nn(),r=l=>{t(l),document.querySelector(".suggestions-list").style.display="block",document.querySelector(".dropdown-search").style.display="block",(l==""||i==!0)&&(s(!1),document.querySelector(".dropdown-search").style.display="none");const d=document.querySelector(".enter-symbol");d&&(d.style.opacity=l?"1":"0")};let n=MC.search(e);const o=l=>{l.preventDefault(),e&&(a(n),s(!0),document.querySelector(".suggestions-list").style.display="none",document.querySelector(".dropdown-search").style.display="none")};return m.jsx(m.Fragment,{children:m.jsxs("div",{className:"search",children:[m.jsxs("form",{className:"search-form",children:[m.jsx("button",{style:{fontSize:"20px",minWidth:"90px",backgroundColor:"white",border:"1px solid #ccc",borderRadius:"4px"},onClick:l=>{o(l)},children:m.jsx(Av,{})}),m.jsxs("div",{className:"input-container",children:[m.jsx("input",{type:"text",placeholder:"Search for plugins",value:e,label:"search",onChange:l=>r(l.target.value)}),m.jsx(Rv,{className:"enter-symbol"})]})]}),m.jsxs("ul",{className:"suggestions-list",children:[n.slice(0,3).map(l=>m.jsxs(m.Fragment,{children:[m.jsx(Bi,{to:`/${l.item.name}`,children:m.jsxs("h3",{className:"suggestion-item",children:[l.item.name," "]},l.item.name)}),m.jsx("ul",{children:l.matches.filter(d=>typeof d.key=="object").slice(0,1).map(d=>m.jsxs(m.Fragment,{children:[m.jsx(Bi,{to:`/${l.item.name}#${d.key[1]}.${d.key[2]}`,children:m.jsxs("li",{className:"suggestion-item",children:[d.key[2]," "]},d.key)}),m.jsx($v,{match_value:d.value})]}))})]})),m.jsx("button",{className:"dropdown-search",onClick:l=>{o(l)},children:" Search"})]})]})})}function WC(){const{searchResults:e,searchQuery:t}=nn();return m.jsxs(m.Fragment,{children:[m.jsxs("h2",{children:["Showing ",e.length," pages matching the search query."]}),e.length===0&&m.jsx("div",{children:m.jsxs("h3",{className:"submenu-entry",style:{textAlign:"center",color:"black"},children:["Can't find what you're looking for?",m.jsx("br",{}),"Join the AiiDA community on Discourse and request a plugin ",m.jsx("a",{href:"https://aiida.discourse.group/new-topic?title=Request%20for%20Plugin...&category=community/plugin-requests",target:"_blank",children:"here."})]})}),e.map(a=>m.jsx(m.Fragment,{children:m.jsxs("div",{className:"submenu-entry",children:[m.jsx(Bi,{to:`/${a.item.name}`,children:m.jsx("h3",{className:"suggestion-item",children:a.item.name},a.item.name)}),m.jsx("ul",{children:a.matches.filter(i=>typeof i.key=="object").map(i=>m.jsx(m.Fragment,{children:zv(i.value,t)[0]!=null&&m.jsxs(m.Fragment,{children:[m.jsx(Bi,{to:`/${a.item.name}#${i.key[1]}.${i.key[2]}`,children:m.jsx("li",{className:"suggestion-item",children:i.key[2]},i.key)}),m.jsx($v,{match_value:i.value})]})}))})]})}))]})}function $v({match_value:e}){const{searchQuery:t}=nn(),[a,i,s]=zv(e,t);return m.jsx(m.Fragment,{children:a!=null&&m.jsxs("p",{children:[a,m.jsx("span",{style:{backgroundColor:"yellow"},children:i}),s,"..."]})})}function zC(e){return Object.keys(e).length===0}function $C(e=null){const t=b.useContext(Rl);return!t||zC(t)?e:t}const UC=Ap();function Mp(e=UC){return $C(e)}function KC({styles:e,themeId:t,defaultTheme:a={}}){const i=Mp(a),s=typeof e=="function"?e(t&&i[t]||i):e;return m.jsx(pv,{styles:s})}const VC=["className","component"];function HC(e={}){const{themeId:t,defaultTheme:a,defaultClassName:i="MuiBox-root",generateClassName:s}=e,r=fv("div",{shouldForwardProp:o=>o!=="theme"&&o!=="sx"&&o!=="as"})(Ml);return b.forwardRef(function(l,d){const c=Mp(a),p=Rp(l),{className:u,component:v="div"}=p,f=U(p,VC);return m.jsx(r,q({as:v,ref:d,className:H(u,s?s(i):i),theme:t&&c[t]||c},f))})}const GC=b.createContext(),Uv=()=>{const e=b.useContext(GC);return e??!1};function on(){const e=Mp(Ep);return e[Ll]||e}const XC=e=>{let t;return e<1?t=5.11916*e**2:t=4.5*Math.log(e+1)+2,(t/100).toFixed(2)},ch=XC,JC=re("MuiBox",["root"]),QC=JC,YC=Cv(),ZC=HC({themeId:Ll,defaultTheme:YC,defaultClassName:QC.root,generateClassName:_p.generate}),e2=ZC;function ln({props:e,states:t,muiFormControl:a}){return t.reduce((i,s)=>(i[s]=e[s],a&&typeof e[s]>"u"&&(i[s]=a[s]),i),{})}const t2=b.createContext(void 0),Lp=t2;function dn(){return b.useContext(Lp)}function a2(e){return pe("MuiFormLabel",e)}const i2=re("MuiFormLabel",["root","colorSecondary","focused","disabled","error","filled","required","asterisk"]),br=i2,s2=["children","className","color","component","disabled","error","filled","focused","required"],r2=e=>{const{classes:t,color:a,focused:i,disabled:s,error:r,filled:n,required:o}=e,l={root:["root",`color${X(a)}`,s&&"disabled",r&&"error",n&&"filled",i&&"focused",o&&"required"],asterisk:["asterisk",r&&"error"]};return fe(l,a2,t)},n2=z("label",{name:"MuiFormLabel",slot:"Root",overridesResolver:({ownerState:e},t)=>q({},t.root,e.color==="secondary"&&t.colorSecondary,e.filled&&t.filled)})(({theme:e,ownerState:t})=>q({color:(e.vars||e).palette.text.secondary},e.typography.body1,{lineHeight:"1.4375em",padding:0,position:"relative",[`&.${br.focused}`]:{color:(e.vars||e).palette[t.color].main},[`&.${br.disabled}`]:{color:(e.vars||e).palette.text.disabled},[`&.${br.error}`]:{color:(e.vars||e).palette.error.main}})),o2=z("span",{name:"MuiFormLabel",slot:"Asterisk",overridesResolver:(e,t)=>t.asterisk})(({theme:e})=>({[`&.${br.error}`]:{color:(e.vars||e).palette.error.main}})),l2=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiFormLabel"}),{children:s,className:r,component:n="label"}=i,o=U(i,s2),l=dn(),d=ln({props:i,muiFormControl:l,states:["color","required","focused","disabled","error","filled"]}),c=q({},i,{color:d.color||"primary",component:n,disabled:d.disabled,error:d.error,filled:d.filled,focused:d.focused,required:d.required}),p=r2(c);return m.jsxs(n2,q({as:n,ownerState:c,className:H(p.root,r),ref:a},o,{children:[s,d.required&&m.jsxs(o2,{ownerState:c,"aria-hidden":!0,className:p.asterisk,children:[" ","*"]})]}))}),d2=l2;function u2(e){return pe("MuiInputLabel",e)}re("MuiInputLabel",["root","focused","disabled","error","required","asterisk","formControl","sizeSmall","shrink","animated","standard","filled","outlined"]);const c2=["disableAnimation","margin","shrink","variant","className"],p2=e=>{const{classes:t,formControl:a,size:i,shrink:s,disableAnimation:r,variant:n,required:o}=e,l={root:["root",a&&"formControl",!r&&"animated",s&&"shrink",i&&i!=="normal"&&`size${X(i)}`,n],asterisk:[o&&"asterisk"]},d=fe(l,u2,t);return q({},t,d)},f2=z(d2,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiInputLabel",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[{[`& .${br.asterisk}`]:t.asterisk},t.root,a.formControl&&t.formControl,a.size==="small"&&t.sizeSmall,a.shrink&&t.shrink,!a.disableAnimation&&t.animated,a.focused&&t.focused,t[a.variant]]}})(({theme:e,ownerState:t})=>q({display:"block",transformOrigin:"top left",whiteSpace:"nowrap",overflow:"hidden",textOverflow:"ellipsis",maxWidth:"100%"},t.formControl&&{position:"absolute",left:0,top:0,transform:"translate(0, 20px) scale(1)"},t.size==="small"&&{transform:"translate(0, 17px) scale(1)"},t.shrink&&{transform:"translate(0, -1.5px) scale(0.75)",transformOrigin:"top left",maxWidth:"133%"},!t.disableAnimation&&{transition:e.transitions.create(["color","transform","max-width"],{duration:e.transitions.duration.shorter,easing:e.transitions.easing.easeOut})},t.variant==="filled"&&q({zIndex:1,pointerEvents:"none",transform:"translate(12px, 16px) scale(1)",maxWidth:"calc(100% - 24px)"},t.size==="small"&&{transform:"translate(12px, 13px) scale(1)"},t.shrink&&q({userSelect:"none",pointerEvents:"auto",transform:"translate(12px, 7px) scale(0.75)",maxWidth:"calc(133% - 24px)"},t.size==="small"&&{transform:"translate(12px, 4px) scale(0.75)"})),t.variant==="outlined"&&q({zIndex:1,pointerEvents:"none",transform:"translate(14px, 16px) scale(1)",maxWidth:"calc(100% - 24px)"},t.size==="small"&&{transform:"translate(14px, 9px) scale(1)"},t.shrink&&{userSelect:"none",pointerEvents:"auto",maxWidth:"calc(133% - 32px)",transform:"translate(14px, -9px) scale(0.75)"}))),m2=b.forwardRef(function(t,a){const i=me({name:"MuiInputLabel",props:t}),{disableAnimation:s=!1,shrink:r,className:n}=i,o=U(i,c2),l=dn();let d=r;typeof d>"u"&&l&&(d=l.filled||l.focused||l.adornedStart);const c=ln({props:i,muiFormControl:l,states:["size","variant","required","focused"]}),p=q({},i,{disableAnimation:s,formControl:l,shrink:d,size:c.size,variant:c.variant,required:c.required,focused:c.focused}),u=p2(p);return m.jsx(f2,q({"data-shrink":d,ownerState:p,ref:a,className:H(u.root,n)},o,{classes:u}))}),h2=m2,g2=b.createContext({}),nc=g2;function oc(e,t){return oc=Object.setPrototypeOf?Object.setPrototypeOf.bind():function(a,i){return a.__proto__=i,a},oc(e,t)}function Kv(e,t){e.prototype=Object.create(t.prototype),e.prototype.constructor=e,oc(e,t)}const ph={disabled:!1},Oo=ea.createContext(null);var y2=function(t){return t.scrollTop},lr="unmounted",bi="exited",wi="entering",Yi="entered",lc="exiting",Oa=function(e){Kv(t,e);function t(i,s){var r;r=e.call(this,i,s)||this;var n=s,o=n&&!n.isMounting?i.enter:i.appear,l;return r.appearStatus=null,i.in?o?(l=bi,r.appearStatus=wi):l=Yi:i.unmountOnExit||i.mountOnEnter?l=lr:l=bi,r.state={status:l},r.nextCallback=null,r}t.getDerivedStateFromProps=function(s,r){var n=s.in;return n&&r.status===lr?{status:bi}:null};var a=t.prototype;return a.componentDidMount=function(){this.updateStatus(!0,this.appearStatus)},a.componentDidUpdate=function(s){var r=null;if(s!==this.props){var n=this.state.status;this.props.in?n!==wi&&n!==Yi&&(r=wi):(n===wi||n===Yi)&&(r=lc)}this.updateStatus(!1,r)},a.componentWillUnmount=function(){this.cancelNextCallback()},a.getTimeouts=function(){var s=this.props.timeout,r,n,o;return r=n=o=s,s!=null&&typeof s!="number"&&(r=s.exit,n=s.enter,o=s.appear!==void 0?s.appear:n),{exit:r,enter:n,appear:o}},a.updateStatus=function(s,r){if(s===void 0&&(s=!1),r!==null)if(this.cancelNextCallback(),r===wi){if(this.props.unmountOnExit||this.props.mountOnEnter){var n=this.props.nodeRef?this.props.nodeRef.current:Sn.findDOMNode(this);n&&y2(n)}this.performEnter(s)}else this.performExit();else this.props.unmountOnExit&&this.state.status===bi&&this.setState({status:lr})},a.performEnter=function(s){var r=this,n=this.props.enter,o=this.context?this.context.isMounting:s,l=this.props.nodeRef?[o]:[Sn.findDOMNode(this),o],d=l[0],c=l[1],p=this.getTimeouts(),u=o?p.appear:p.enter;if(!s&&!n||ph.disabled){this.safeSetState({status:Yi},function(){r.props.onEntered(d)});return}this.props.onEnter(d,c),this.safeSetState({status:wi},function(){r.props.onEntering(d,c),r.onTransitionEnd(u,function(){r.safeSetState({status:Yi},function(){r.props.onEntered(d,c)})})})},a.performExit=function(){var s=this,r=this.props.exit,n=this.getTimeouts(),o=this.props.nodeRef?void 0:Sn.findDOMNode(this);if(!r||ph.disabled){this.safeSetState({status:bi},function(){s.props.onExited(o)});return}this.props.onExit(o),this.safeSetState({status:lc},function(){s.props.onExiting(o),s.onTransitionEnd(n.exit,function(){s.safeSetState({status:bi},function(){s.props.onExited(o)})})})},a.cancelNextCallback=function(){this.nextCallback!==null&&(this.nextCallback.cancel(),this.nextCallback=null)},a.safeSetState=function(s,r){r=this.setNextCallback(r),this.setState(s,r)},a.setNextCallback=function(s){var r=this,n=!0;return this.nextCallback=function(o){n&&(n=!1,r.nextCallback=null,s(o))},this.nextCallback.cancel=function(){n=!1},this.nextCallback},a.onTransitionEnd=function(s,r){this.setNextCallback(r);var n=this.props.nodeRef?this.props.nodeRef.current:Sn.findDOMNode(this),o=s==null&&!this.props.addEndListener;if(!n||o){setTimeout(this.nextCallback,0);return}if(this.props.addEndListener){var l=this.props.nodeRef?[this.nextCallback]:[n,this.nextCallback],d=l[0],c=l[1];this.props.addEndListener(d,c)}s!=null&&setTimeout(this.nextCallback,s)},a.render=function(){var s=this.state.status;if(s===lr)return null;var r=this.props,n=r.children;r.in,r.mountOnEnter,r.unmountOnExit,r.appear,r.enter,r.exit,r.timeout,r.addEndListener,r.onEnter,r.onEntering,r.onEntered,r.onExit,r.onExiting,r.onExited,r.nodeRef;var o=U(r,["children","in","mountOnEnter","unmountOnExit","appear","enter","exit","timeout","addEndListener","onEnter","onEntering","onEntered","onExit","onExiting","onExited","nodeRef"]);return ea.createElement(Oo.Provider,{value:null},typeof n=="function"?n(s,o):ea.cloneElement(ea.Children.only(n),o))},t}(ea.Component);Oa.contextType=Oo;Oa.propTypes={};function Qi(){}Oa.defaultProps={in:!1,mountOnEnter:!1,unmountOnExit:!1,appear:!1,enter:!0,exit:!0,onEnter:Qi,onEntering:Qi,onEntered:Qi,onExit:Qi,onExiting:Qi,onExited:Qi};Oa.UNMOUNTED=lr;Oa.EXITED=bi;Oa.ENTERING=wi;Oa.ENTERED=Yi;Oa.EXITING=lc;const Wp=Oa;function _2(e){if(e===void 0)throw new ReferenceError("this hasn't been initialised - super() hasn't been called");return e}function zp(e,t){var a=function(r){return t&&b.isValidElement(r)?t(r):r},i=Object.create(null);return e&&b.Children.map(e,function(s){return s}).forEach(function(s){i[s.key]=a(s)}),i}function v2(e,t){e=e||{},t=t||{};function a(c){return c in t?t[c]:e[c]}var i=Object.create(null),s=[];for(var r in e)r in t?s.length&&(i[r]=s,s=[]):s.push(r);var n,o={};for(var l in t){if(i[l])for(n=0;n{if(!o&&l!=null){const h=setTimeout(l,d);return()=>{clearTimeout(h)}}},[l,o,d]),m.jsx("span",{className:u,style:v,children:m.jsx("span",{className:f})})}const x2=re("MuiTouchRipple",["root","ripple","rippleVisible","ripplePulsate","child","childLeaving","childPulsate"]),Ot=x2,C2=["center","classes","className"];let Wl=e=>e,fh,mh,hh,gh;const dc=550,S2=80,P2=Il(fh||(fh=Wl` + 0% { + transform: scale(0); + opacity: 0.1; + } + + 100% { + transform: scale(1); + opacity: 0.3; + } +`)),A2=Il(mh||(mh=Wl` + 0% { + opacity: 1; + } + + 100% { + opacity: 0; + } +`)),R2=Il(hh||(hh=Wl` + 0% { + transform: scale(1); + } + + 50% { + transform: scale(0.92); + } + + 100% { + transform: scale(1); + } +`)),I2=z("span",{name:"MuiTouchRipple",slot:"Root"})({overflow:"hidden",pointerEvents:"none",position:"absolute",zIndex:0,top:0,right:0,bottom:0,left:0,borderRadius:"inherit"}),N2=z(D2,{name:"MuiTouchRipple",slot:"Ripple"})(gh||(gh=Wl` + opacity: 0; + position: absolute; + + &.${0} { + opacity: 0.3; + transform: scale(1); + animation-name: ${0}; + animation-duration: ${0}ms; + animation-timing-function: ${0}; + } + + &.${0} { + animation-duration: ${0}ms; + } + + & .${0} { + opacity: 1; + display: block; + width: 100%; + height: 100%; + border-radius: 50%; + background-color: currentColor; + } + + & .${0} { + opacity: 0; + animation-name: ${0}; + animation-duration: ${0}ms; + animation-timing-function: ${0}; + } + + & .${0} { + position: absolute; + /* @noflip */ + left: 0px; + top: 0; + animation-name: ${0}; + animation-duration: 2500ms; + animation-timing-function: ${0}; + animation-iteration-count: infinite; + animation-delay: 200ms; + } +`),Ot.rippleVisible,P2,dc,({theme:e})=>e.transitions.easing.easeInOut,Ot.ripplePulsate,({theme:e})=>e.transitions.duration.shorter,Ot.child,Ot.childLeaving,A2,dc,({theme:e})=>e.transitions.easing.easeInOut,Ot.childPulsate,R2,({theme:e})=>e.transitions.easing.easeInOut),j2=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiTouchRipple"}),{center:s=!1,classes:r={},className:n}=i,o=U(i,C2),[l,d]=b.useState([]),c=b.useRef(0),p=b.useRef(null);b.useEffect(()=>{p.current&&(p.current(),p.current=null)},[l]);const u=b.useRef(!1),v=B_(),f=b.useRef(null),h=b.useRef(null),T=b.useCallback(w=>{const{pulsate:D,rippleX:x,rippleY:k,rippleSize:S,cb:A}=w;d(P=>[...P,m.jsx(N2,{classes:{ripple:H(r.ripple,Ot.ripple),rippleVisible:H(r.rippleVisible,Ot.rippleVisible),ripplePulsate:H(r.ripplePulsate,Ot.ripplePulsate),child:H(r.child,Ot.child),childLeaving:H(r.childLeaving,Ot.childLeaving),childPulsate:H(r.childPulsate,Ot.childPulsate)},timeout:dc,pulsate:D,rippleX:x,rippleY:k,rippleSize:S},c.current)]),c.current+=1,p.current=A},[r]),y=b.useCallback((w={},D={},x=()=>{})=>{const{pulsate:k=!1,center:S=s||D.pulsate,fakeElement:A=!1}=D;if((w==null?void 0:w.type)==="mousedown"&&u.current){u.current=!1;return}(w==null?void 0:w.type)==="touchstart"&&(u.current=!0);const P=A?null:h.current,j=P?P.getBoundingClientRect():{width:0,height:0,left:0,top:0};let E,F,I;if(S||w===void 0||w.clientX===0&&w.clientY===0||!w.clientX&&!w.touches)E=Math.round(j.width/2),F=Math.round(j.height/2);else{const{clientX:B,clientY:O}=w.touches&&w.touches.length>0?w.touches[0]:w;E=Math.round(B-j.left),F=Math.round(O-j.top)}if(S)I=Math.sqrt((2*j.width**2+j.height**2)/3),I%2===0&&(I+=1);else{const B=Math.max(Math.abs((P?P.clientWidth:0)-E),E)*2+2,O=Math.max(Math.abs((P?P.clientHeight:0)-F),F)*2+2;I=Math.sqrt(B**2+O**2)}w!=null&&w.touches?f.current===null&&(f.current=()=>{T({pulsate:k,rippleX:E,rippleY:F,rippleSize:I,cb:x})},v.start(S2,()=>{f.current&&(f.current(),f.current=null)})):T({pulsate:k,rippleX:E,rippleY:F,rippleSize:I,cb:x})},[s,T,v]),g=b.useCallback(()=>{y({},{pulsate:!0})},[y]),_=b.useCallback((w,D)=>{if(v.clear(),(w==null?void 0:w.type)==="touchend"&&f.current){f.current(),f.current=null,v.start(0,()=>{_(w,D)});return}f.current=null,d(x=>x.length>0?x.slice(1):x),p.current=D},[v]);return b.useImperativeHandle(a,()=>({pulsate:g,start:y,stop:_}),[g,y,_]),m.jsx(I2,q({className:H(Ot.root,r.root,n),ref:h},o,{children:m.jsx(q2,{component:null,exit:!0,children:l})}))}),E2=j2;function F2(e){return pe("MuiButtonBase",e)}const B2=re("MuiButtonBase",["root","disabled","focusVisible"]),O2=B2,M2=["action","centerRipple","children","className","component","disabled","disableRipple","disableTouchRipple","focusRipple","focusVisibleClassName","LinkComponent","onBlur","onClick","onContextMenu","onDragLeave","onFocus","onFocusVisible","onKeyDown","onKeyUp","onMouseDown","onMouseLeave","onMouseUp","onTouchEnd","onTouchMove","onTouchStart","tabIndex","TouchRippleProps","touchRippleRef","type"],L2=e=>{const{disabled:t,focusVisible:a,focusVisibleClassName:i,classes:s}=e,n=fe({root:["root",t&&"disabled",a&&"focusVisible"]},F2,s);return a&&i&&(n.root+=` ${i}`),n},W2=z("button",{name:"MuiButtonBase",slot:"Root",overridesResolver:(e,t)=>t.root})({display:"inline-flex",alignItems:"center",justifyContent:"center",position:"relative",boxSizing:"border-box",WebkitTapHighlightColor:"transparent",backgroundColor:"transparent",outline:0,border:0,margin:0,borderRadius:0,padding:0,cursor:"pointer",userSelect:"none",verticalAlign:"middle",MozAppearance:"none",WebkitAppearance:"none",textDecoration:"none",color:"inherit","&::-moz-focus-inner":{borderStyle:"none"},[`&.${O2.disabled}`]:{pointerEvents:"none",cursor:"default"},"@media print":{colorAdjust:"exact"}}),z2=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiButtonBase"}),{action:s,centerRipple:r=!1,children:n,className:o,component:l="button",disabled:d=!1,disableRipple:c=!1,disableTouchRipple:p=!1,focusRipple:u=!1,LinkComponent:v="a",onBlur:f,onClick:h,onContextMenu:T,onDragLeave:y,onFocus:g,onFocusVisible:_,onKeyDown:w,onKeyUp:D,onMouseDown:x,onMouseLeave:k,onMouseUp:S,onTouchEnd:A,onTouchMove:P,onTouchStart:j,tabIndex:E=0,TouchRippleProps:F,touchRippleRef:I,type:B}=i,O=U(i,M2),W=b.useRef(null),C=b.useRef(null),N=Ke(C,I),{isFocusVisibleRef:L,onFocus:te,onBlur:J,ref:Ne}=O_(),[ee,be]=b.useState(!1);d&&ee&&be(!1),b.useImperativeHandle(s,()=>({focusVisible:()=>{be(!0),W.current.focus()}}),[]);const[oe,He]=b.useState(!1);b.useEffect(()=>{He(!0)},[]);const Nt=oe&&!c&&!d;b.useEffect(()=>{ee&&u&&!c&&oe&&C.current.pulsate()},[c,u,ee,oe]);function je(V,ba,Ls=p){return Si(Ws=>(ba&&ba(Ws),!Ls&&C.current&&C.current[V](Ws),!0))}const ut=je("start",x),ne=je("stop",T),De=je("stop",y),Z=je("stop",S),ce=je("stop",V=>{ee&&V.preventDefault(),k&&k(V)}),we=je("start",j),Ma=je("stop",A),jt=je("stop",P),Et=je("stop",V=>{J(V),L.current===!1&&be(!1),f&&f(V)},!1),Xt=Si(V=>{W.current||(W.current=V.currentTarget),te(V),L.current===!0&&(be(!0),_&&_(V)),g&&g(V)}),Ft=()=>{const V=W.current;return l&&l!=="button"&&!(V.tagName==="A"&&V.href)},xe=b.useRef(!1),_a=Si(V=>{u&&!xe.current&&ee&&C.current&&V.key===" "&&(xe.current=!0,C.current.stop(V,()=>{C.current.start(V)})),V.target===V.currentTarget&&Ft()&&V.key===" "&&V.preventDefault(),w&&w(V),V.target===V.currentTarget&&Ft()&&V.key==="Enter"&&!d&&(V.preventDefault(),h&&h(V))}),ct=Si(V=>{u&&V.key===" "&&C.current&&ee&&!V.defaultPrevented&&(xe.current=!1,C.current.stop(V,()=>{C.current.pulsate(V)})),D&&D(V),h&&V.target===V.currentTarget&&Ft()&&V.key===" "&&!V.defaultPrevented&&h(V)});let Te=l;Te==="button"&&(O.href||O.to)&&(Te=v);const ra={};Te==="button"?(ra.type=B===void 0?"button":B,ra.disabled=d):(!O.href&&!O.to&&(ra.role="button"),d&&(ra["aria-disabled"]=d));const La=Ke(a,Ne,W),va=q({},i,{centerRipple:r,component:l,disabled:d,disableRipple:c,disableTouchRipple:p,focusRipple:u,tabIndex:E,focusVisible:ee}),he=L2(va);return m.jsxs(W2,q({as:Te,className:H(he.root,o),ownerState:va,onBlur:Et,onClick:h,onContextMenu:ne,onFocus:Xt,onKeyDown:_a,onKeyUp:ct,onMouseDown:ut,onMouseLeave:ce,onMouseUp:Z,onDragLeave:De,onTouchEnd:Ma,onTouchMove:jt,onTouchStart:we,ref:La,tabIndex:d?-1:E,type:B},ra,O,{children:[n,Nt?m.jsx(E2,q({ref:N,center:r},F)):null]}))}),Vv=z2;function $2(e){return pe("MuiDivider",e)}const U2=re("MuiDivider",["root","absolute","fullWidth","inset","middle","flexItem","light","vertical","withChildren","withChildrenVertical","textAlignRight","textAlignLeft","wrapper","wrapperVertical"]),yh=U2,K2=["absolute","children","className","component","flexItem","light","orientation","role","textAlign","variant"],V2=e=>{const{absolute:t,children:a,classes:i,flexItem:s,light:r,orientation:n,textAlign:o,variant:l}=e;return fe({root:["root",t&&"absolute",l,r&&"light",n==="vertical"&&"vertical",s&&"flexItem",a&&"withChildren",a&&n==="vertical"&&"withChildrenVertical",o==="right"&&n!=="vertical"&&"textAlignRight",o==="left"&&n!=="vertical"&&"textAlignLeft"],wrapper:["wrapper",n==="vertical"&&"wrapperVertical"]},$2,i)},H2=z("div",{name:"MuiDivider",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,a.absolute&&t.absolute,t[a.variant],a.light&&t.light,a.orientation==="vertical"&&t.vertical,a.flexItem&&t.flexItem,a.children&&t.withChildren,a.children&&a.orientation==="vertical"&&t.withChildrenVertical,a.textAlign==="right"&&a.orientation!=="vertical"&&t.textAlignRight,a.textAlign==="left"&&a.orientation!=="vertical"&&t.textAlignLeft]}})(({theme:e,ownerState:t})=>q({margin:0,flexShrink:0,borderWidth:0,borderStyle:"solid",borderColor:(e.vars||e).palette.divider,borderBottomWidth:"thin"},t.absolute&&{position:"absolute",bottom:0,left:0,width:"100%"},t.light&&{borderColor:e.vars?`rgba(${e.vars.palette.dividerChannel} / 0.08)`:xa(e.palette.divider,.08)},t.variant==="inset"&&{marginLeft:72},t.variant==="middle"&&t.orientation==="horizontal"&&{marginLeft:e.spacing(2),marginRight:e.spacing(2)},t.variant==="middle"&&t.orientation==="vertical"&&{marginTop:e.spacing(1),marginBottom:e.spacing(1)},t.orientation==="vertical"&&{height:"100%",borderBottomWidth:0,borderRightWidth:"thin"},t.flexItem&&{alignSelf:"stretch",height:"auto"}),({ownerState:e})=>q({},e.children&&{display:"flex",whiteSpace:"nowrap",textAlign:"center",border:0,borderTopStyle:"solid",borderLeftStyle:"solid","&::before, &::after":{content:'""',alignSelf:"center"}}),({theme:e,ownerState:t})=>q({},t.children&&t.orientation!=="vertical"&&{"&::before, &::after":{width:"100%",borderTop:`thin solid ${(e.vars||e).palette.divider}`,borderTopStyle:"inherit"}}),({theme:e,ownerState:t})=>q({},t.children&&t.orientation==="vertical"&&{flexDirection:"column","&::before, &::after":{height:"100%",borderLeft:`thin solid ${(e.vars||e).palette.divider}`,borderLeftStyle:"inherit"}}),({ownerState:e})=>q({},e.textAlign==="right"&&e.orientation!=="vertical"&&{"&::before":{width:"90%"},"&::after":{width:"10%"}},e.textAlign==="left"&&e.orientation!=="vertical"&&{"&::before":{width:"10%"},"&::after":{width:"90%"}})),G2=z("span",{name:"MuiDivider",slot:"Wrapper",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.wrapper,a.orientation==="vertical"&&t.wrapperVertical]}})(({theme:e,ownerState:t})=>q({display:"inline-block",paddingLeft:`calc(${e.spacing(1)} * 1.2)`,paddingRight:`calc(${e.spacing(1)} * 1.2)`},t.orientation==="vertical"&&{paddingTop:`calc(${e.spacing(1)} * 1.2)`,paddingBottom:`calc(${e.spacing(1)} * 1.2)`})),Hv=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiDivider"}),{absolute:s=!1,children:r,className:n,component:o=r?"div":"hr",flexItem:l=!1,light:d=!1,orientation:c="horizontal",role:p=o!=="hr"?"separator":void 0,textAlign:u="center",variant:v="fullWidth"}=i,f=U(i,K2),h=q({},i,{absolute:s,component:o,flexItem:l,light:d,orientation:c,role:p,textAlign:u,variant:v}),T=V2(h);return m.jsx(H2,q({as:o,className:H(T.root,n),role:p,ref:a,ownerState:h},f,{children:r?m.jsx(G2,{className:T.wrapper,ownerState:h,children:r}):null}))});Hv.muiSkipListHighlight=!0;const _h=Hv,X2=re("MuiListItemIcon",["root","alignItemsFlexStart"]),vh=X2;function J2(e){return pe("MuiTypography",e)}re("MuiTypography",["root","h1","h2","h3","h4","h5","h6","subtitle1","subtitle2","body1","body2","inherit","button","caption","overline","alignLeft","alignRight","alignCenter","alignJustify","noWrap","gutterBottom","paragraph"]);const Q2=["align","className","component","gutterBottom","noWrap","paragraph","variant","variantMapping"],Y2=e=>{const{align:t,gutterBottom:a,noWrap:i,paragraph:s,variant:r,classes:n}=e,o={root:["root",r,e.align!=="inherit"&&`align${X(t)}`,a&&"gutterBottom",i&&"noWrap",s&&"paragraph"]};return fe(o,J2,n)},Z2=z("span",{name:"MuiTypography",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,a.variant&&t[a.variant],a.align!=="inherit"&&t[`align${X(a.align)}`],a.noWrap&&t.noWrap,a.gutterBottom&&t.gutterBottom,a.paragraph&&t.paragraph]}})(({theme:e,ownerState:t})=>q({margin:0},t.variant==="inherit"&&{font:"inherit"},t.variant!=="inherit"&&e.typography[t.variant],t.align!=="inherit"&&{textAlign:t.align},t.noWrap&&{overflow:"hidden",textOverflow:"ellipsis",whiteSpace:"nowrap"},t.gutterBottom&&{marginBottom:"0.35em"},t.paragraph&&{marginBottom:16})),bh={h1:"h1",h2:"h2",h3:"h3",h4:"h4",h5:"h5",h6:"h6",subtitle1:"h6",subtitle2:"h6",body1:"p",body2:"p",inherit:"p"},eS={primary:"primary.main",textPrimary:"text.primary",secondary:"secondary.main",textSecondary:"text.secondary",error:"error.main"},tS=e=>eS[e]||e,aS=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiTypography"}),s=tS(i.color),r=Rp(q({},i,{color:s})),{align:n="inherit",className:o,component:l,gutterBottom:d=!1,noWrap:c=!1,paragraph:p=!1,variant:u="body1",variantMapping:v=bh}=r,f=U(r,Q2),h=q({},r,{align:n,color:s,className:o,component:l,gutterBottom:d,noWrap:c,paragraph:p,variant:u,variantMapping:v}),T=l||(p?"p":v[u]||bh[u])||"span",y=Y2(h);return m.jsx(Z2,q({as:T,ref:a,ownerState:h,className:H(y.root,o)},f))}),Gv=aS,iS=re("MuiListItemText",["root","multiline","dense","inset","primary","secondary"]),wh=iS;function sS(e){return pe("MuiMenuItem",e)}const rS=re("MuiMenuItem",["root","focusVisible","dense","disabled","divider","gutters","selected"]),Ys=rS,nS=["autoFocus","component","dense","divider","disableGutters","focusVisibleClassName","role","tabIndex","className"],oS=(e,t)=>{const{ownerState:a}=e;return[t.root,a.dense&&t.dense,a.divider&&t.divider,!a.disableGutters&&t.gutters]},lS=e=>{const{disabled:t,dense:a,divider:i,disableGutters:s,selected:r,classes:n}=e,l=fe({root:["root",a&&"dense",t&&"disabled",!s&&"gutters",i&&"divider",r&&"selected"]},sS,n);return q({},n,l)},dS=z(Vv,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiMenuItem",slot:"Root",overridesResolver:oS})(({theme:e,ownerState:t})=>q({},e.typography.body1,{display:"flex",justifyContent:"flex-start",alignItems:"center",position:"relative",textDecoration:"none",minHeight:48,paddingTop:6,paddingBottom:6,boxSizing:"border-box",whiteSpace:"nowrap"},!t.disableGutters&&{paddingLeft:16,paddingRight:16},t.divider&&{borderBottom:`1px solid ${(e.vars||e).palette.divider}`,backgroundClip:"padding-box"},{"&:hover":{textDecoration:"none",backgroundColor:(e.vars||e).palette.action.hover,"@media (hover: none)":{backgroundColor:"transparent"}},[`&.${Ys.selected}`]:{backgroundColor:e.vars?`rgba(${e.vars.palette.primary.mainChannel} / ${e.vars.palette.action.selectedOpacity})`:xa(e.palette.primary.main,e.palette.action.selectedOpacity),[`&.${Ys.focusVisible}`]:{backgroundColor:e.vars?`rgba(${e.vars.palette.primary.mainChannel} / calc(${e.vars.palette.action.selectedOpacity} + ${e.vars.palette.action.focusOpacity}))`:xa(e.palette.primary.main,e.palette.action.selectedOpacity+e.palette.action.focusOpacity)}},[`&.${Ys.selected}:hover`]:{backgroundColor:e.vars?`rgba(${e.vars.palette.primary.mainChannel} / calc(${e.vars.palette.action.selectedOpacity} + ${e.vars.palette.action.hoverOpacity}))`:xa(e.palette.primary.main,e.palette.action.selectedOpacity+e.palette.action.hoverOpacity),"@media (hover: none)":{backgroundColor:e.vars?`rgba(${e.vars.palette.primary.mainChannel} / ${e.vars.palette.action.selectedOpacity})`:xa(e.palette.primary.main,e.palette.action.selectedOpacity)}},[`&.${Ys.focusVisible}`]:{backgroundColor:(e.vars||e).palette.action.focus},[`&.${Ys.disabled}`]:{opacity:(e.vars||e).palette.action.disabledOpacity},[`& + .${yh.root}`]:{marginTop:e.spacing(1),marginBottom:e.spacing(1)},[`& + .${yh.inset}`]:{marginLeft:52},[`& .${wh.root}`]:{marginTop:0,marginBottom:0},[`& .${wh.inset}`]:{paddingLeft:36},[`& .${vh.root}`]:{minWidth:36}},!t.dense&&{[e.breakpoints.up("sm")]:{minHeight:"auto"}},t.dense&&q({minHeight:32,paddingTop:4,paddingBottom:4},e.typography.body2,{[`& .${vh.root} svg`]:{fontSize:"1.25rem"}}))),uS=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiMenuItem"}),{autoFocus:s=!1,component:r="li",dense:n=!1,divider:o=!1,disableGutters:l=!1,focusVisibleClassName:d,role:c="menuitem",tabIndex:p,className:u}=i,v=U(i,nS),f=b.useContext(nc),h=b.useMemo(()=>({dense:n||f.dense||!1,disableGutters:l}),[f.dense,n,l]),T=b.useRef(null);Ia(()=>{s&&T.current&&T.current.focus()},[s]);const y=q({},i,{dense:h.dense,divider:o,disableGutters:l}),g=lS(i),_=Ke(T,a);let w;return i.disabled||(w=p!==void 0?p:-1),m.jsx(nc.Provider,{value:h,children:m.jsx(dS,q({ref:_,role:c,tabIndex:w,component:r,focusVisibleClassName:H(g.focusVisible,d),className:H(g.root,u)},v,{ownerState:y,classes:g}))})}),jd=uS;function Th(e){return e!=null&&!(Array.isArray(e)&&e.length===0)}function Mo(e,t=!1){return e&&(Th(e.value)&&e.value!==""||t&&Th(e.defaultValue)&&e.defaultValue!=="")}function cS(e){return e.startAdornment}function pS(e){return pe("MuiFormControl",e)}re("MuiFormControl",["root","marginNone","marginNormal","marginDense","fullWidth","disabled"]);const fS=["children","className","color","component","disabled","error","focused","fullWidth","hiddenLabel","margin","required","size","variant"],mS=e=>{const{classes:t,margin:a,fullWidth:i}=e,s={root:["root",a!=="none"&&`margin${X(a)}`,i&&"fullWidth"]};return fe(s,pS,t)},hS=z("div",{name:"MuiFormControl",slot:"Root",overridesResolver:({ownerState:e},t)=>q({},t.root,t[`margin${X(e.margin)}`],e.fullWidth&&t.fullWidth)})(({ownerState:e})=>q({display:"inline-flex",flexDirection:"column",position:"relative",minWidth:0,padding:0,margin:0,border:0,verticalAlign:"top"},e.margin==="normal"&&{marginTop:16,marginBottom:8},e.margin==="dense"&&{marginTop:8,marginBottom:4},e.fullWidth&&{width:"100%"})),gS=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiFormControl"}),{children:s,className:r,color:n="primary",component:o="div",disabled:l=!1,error:d=!1,focused:c,fullWidth:p=!1,hiddenLabel:u=!1,margin:v="none",required:f=!1,size:h="medium",variant:T="outlined"}=i,y=U(i,fS),g=q({},i,{color:n,component:o,disabled:l,error:d,fullWidth:p,hiddenLabel:u,margin:v,required:f,size:h,variant:T}),_=mS(g),[w,D]=b.useState(()=>{let F=!1;return s&&b.Children.forEach(s,I=>{if(!Qn(I,["Input","Select"]))return;const B=Qn(I,["Select"])?I.props.input:I;B&&cS(B.props)&&(F=!0)}),F}),[x,k]=b.useState(()=>{let F=!1;return s&&b.Children.forEach(s,I=>{Qn(I,["Input","Select"])&&(Mo(I.props,!0)||Mo(I.props.inputProps,!0))&&(F=!0)}),F}),[S,A]=b.useState(!1);l&&S&&A(!1);const P=c!==void 0&&!l?c:S;let j;const E=b.useMemo(()=>({adornedStart:w,setAdornedStart:D,color:n,disabled:l,error:d,filled:x,focused:P,fullWidth:p,hiddenLabel:u,size:h,onBlur:()=>{A(!1)},onEmpty:()=>{k(!1)},onFilled:()=>{k(!0)},onFocus:()=>{A(!0)},registerEffect:j,required:f,variant:T}),[w,n,l,d,x,P,p,u,j,f,h,T]);return m.jsx(Lp.Provider,{value:E,children:m.jsx(hS,q({as:o,ownerState:g,className:H(_.root,r),ref:a},y,{children:s}))})}),yS=gS;function _S(e){return pe("MuiList",e)}re("MuiList",["root","padding","dense","subheader"]);const vS=["children","className","component","dense","disablePadding","subheader"],bS=e=>{const{classes:t,disablePadding:a,dense:i,subheader:s}=e;return fe({root:["root",!a&&"padding",i&&"dense",s&&"subheader"]},_S,t)},wS=z("ul",{name:"MuiList",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,!a.disablePadding&&t.padding,a.dense&&t.dense,a.subheader&&t.subheader]}})(({ownerState:e})=>q({listStyle:"none",margin:0,padding:0,position:"relative"},!e.disablePadding&&{paddingTop:8,paddingBottom:8},e.subheader&&{paddingTop:0})),TS=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiList"}),{children:s,className:r,component:n="ul",dense:o=!1,disablePadding:l=!1,subheader:d}=i,c=U(i,vS),p=b.useMemo(()=>({dense:o}),[o]),u=q({},i,{component:n,dense:o,disablePadding:l}),v=bS(u);return m.jsx(nc.Provider,{value:p,children:m.jsxs(wS,q({as:n,className:H(v.root,r),ref:a,ownerState:u},c,{children:[d,s]}))})}),kS=TS,qS=["actions","autoFocus","autoFocusItem","children","className","disabledItemsFocusable","disableListWrap","onKeyDown","variant"];function Ed(e,t,a){return e===t?e.firstChild:t&&t.nextElementSibling?t.nextElementSibling:a?null:e.firstChild}function kh(e,t,a){return e===t?a?e.firstChild:e.lastChild:t&&t.previousElementSibling?t.previousElementSibling:a?null:e.lastChild}function Xv(e,t){if(t===void 0)return!0;let a=e.innerText;return a===void 0&&(a=e.textContent),a=a.trim().toLowerCase(),a.length===0?!1:t.repeating?a[0]===t.keys[0]:a.indexOf(t.keys.join(""))===0}function Zs(e,t,a,i,s,r){let n=!1,o=s(e,t,t?a:!1);for(;o;){if(o===e.firstChild){if(n)return!1;n=!0}const l=i?!1:o.disabled||o.getAttribute("aria-disabled")==="true";if(!o.hasAttribute("tabindex")||!Xv(o,r)||l)o=s(e,o,a);else return o.focus(),!0}return!1}const DS=b.forwardRef(function(t,a){const{actions:i,autoFocus:s=!1,autoFocusItem:r=!1,children:n,className:o,disabledItemsFocusable:l=!1,disableListWrap:d=!1,onKeyDown:c,variant:p="selectedMenu"}=t,u=U(t,qS),v=b.useRef(null),f=b.useRef({keys:[],repeating:!0,previousKeyMatched:!0,lastTime:null});Ia(()=>{s&&v.current.focus()},[s]),b.useImperativeHandle(i,()=>({adjustStyleForScrollbar:(_,{direction:w})=>{const D=!v.current.style.width;if(_.clientHeight{const w=v.current,D=_.key,x=_t(w).activeElement;if(D==="ArrowDown")_.preventDefault(),Zs(w,x,d,l,Ed);else if(D==="ArrowUp")_.preventDefault(),Zs(w,x,d,l,kh);else if(D==="Home")_.preventDefault(),Zs(w,null,d,l,Ed);else if(D==="End")_.preventDefault(),Zs(w,null,d,l,kh);else if(D.length===1){const k=f.current,S=D.toLowerCase(),A=performance.now();k.keys.length>0&&(A-k.lastTime>500?(k.keys=[],k.repeating=!0,k.previousKeyMatched=!0):k.repeating&&S!==k.keys[0]&&(k.repeating=!1)),k.lastTime=A,k.keys.push(S);const P=x&&!k.repeating&&Xv(x,k);k.previousKeyMatched&&(P||Zs(w,x,!1,l,Ed,k))?_.preventDefault():k.previousKeyMatched=!1}c&&c(_)},T=Ke(v,a);let y=-1;b.Children.forEach(n,(_,w)=>{if(!b.isValidElement(_)){y===w&&(y+=1,y>=n.length&&(y=-1));return}_.props.disabled||(p==="selectedMenu"&&_.props.selected||y===-1)&&(y=w),y===w&&(_.props.disabled||_.props.muiSkipListHighlight||_.type.muiSkipListHighlight)&&(y+=1,y>=n.length&&(y=-1))});const g=b.Children.map(n,(_,w)=>{if(w===y){const D={};return r&&(D.autoFocus=!0),_.props.tabIndex===void 0&&p==="selectedMenu"&&(D.tabIndex=0),b.cloneElement(_,D)}return _});return m.jsx(kS,q({role:"menu",ref:T,className:o,onKeyDown:h,tabIndex:s?0:-1},u,{children:g}))}),xS=DS,Up=e=>e.scrollTop;function Ps(e,t){var a,i;const{timeout:s,easing:r,style:n={}}=e;return{duration:(a=n.transitionDuration)!=null?a:typeof s=="number"?s:s[t.mode]||0,easing:(i=n.transitionTimingFunction)!=null?i:typeof r=="object"?r[t.mode]:r,delay:n.transitionDelay}}const CS=["addEndListener","appear","children","easing","in","onEnter","onEntered","onEntering","onExit","onExited","onExiting","style","timeout","TransitionComponent"];function uc(e){return`scale(${e}, ${e**2})`}const SS={entering:{opacity:1,transform:uc(1)},entered:{opacity:1,transform:"none"}},Fd=typeof navigator<"u"&&/^((?!chrome|android).)*(safari|mobile)/i.test(navigator.userAgent)&&/(os |version\/)15(.|_)4/i.test(navigator.userAgent),Jv=b.forwardRef(function(t,a){const{addEndListener:i,appear:s=!0,children:r,easing:n,in:o,onEnter:l,onEntered:d,onEntering:c,onExit:p,onExited:u,onExiting:v,style:f,timeout:h="auto",TransitionComponent:T=Wp}=t,y=U(t,CS),g=B_(),_=b.useRef(),w=on(),D=b.useRef(null),x=Ke(D,r.ref,a),k=B=>O=>{if(B){const W=D.current;O===void 0?B(W):B(W,O)}},S=k(c),A=k((B,O)=>{Up(B);const{duration:W,delay:C,easing:N}=Ps({style:f,timeout:h,easing:n},{mode:"enter"});let L;h==="auto"?(L=w.transitions.getAutoHeightDuration(B.clientHeight),_.current=L):L=W,B.style.transition=[w.transitions.create("opacity",{duration:L,delay:C}),w.transitions.create("transform",{duration:Fd?L:L*.666,delay:C,easing:N})].join(","),l&&l(B,O)}),P=k(d),j=k(v),E=k(B=>{const{duration:O,delay:W,easing:C}=Ps({style:f,timeout:h,easing:n},{mode:"exit"});let N;h==="auto"?(N=w.transitions.getAutoHeightDuration(B.clientHeight),_.current=N):N=O,B.style.transition=[w.transitions.create("opacity",{duration:N,delay:W}),w.transitions.create("transform",{duration:Fd?N:N*.666,delay:Fd?W:W||N*.333,easing:C})].join(","),B.style.opacity=0,B.style.transform=uc(.75),p&&p(B)}),F=k(u),I=B=>{h==="auto"&&g.start(_.current||0,B),i&&i(D.current,B)};return m.jsx(T,q({appear:s,in:o,nodeRef:D,onEnter:A,onEntered:P,onEntering:S,onExit:E,onExited:F,onExiting:j,addEndListener:I,timeout:h==="auto"?null:h},y,{children:(B,O)=>b.cloneElement(r,q({style:q({opacity:0,transform:uc(.75),visibility:B==="exited"&&!o?"hidden":void 0},SS[B],f,r.props.style),ref:x},O))}))});Jv.muiSupportAuto=!0;const PS=Jv;function AS(e){const t=_t(e);return t.body===e?ha(e).innerWidth>t.documentElement.clientWidth:e.scrollHeight>e.clientHeight}function wr(e,t){t?e.setAttribute("aria-hidden","true"):e.removeAttribute("aria-hidden")}function qh(e){return parseInt(ha(e).getComputedStyle(e).paddingRight,10)||0}function RS(e){const a=["TEMPLATE","SCRIPT","STYLE","LINK","MAP","META","NOSCRIPT","PICTURE","COL","COLGROUP","PARAM","SLOT","SOURCE","TRACK"].indexOf(e.tagName)!==-1,i=e.tagName==="INPUT"&&e.getAttribute("type")==="hidden";return a||i}function Dh(e,t,a,i,s){const r=[t,a,...i];[].forEach.call(e.children,n=>{const o=r.indexOf(n)===-1,l=!RS(n);o&&l&&wr(n,s)})}function Bd(e,t){let a=-1;return e.some((i,s)=>t(i)?(a=s,!0):!1),a}function IS(e,t){const a=[],i=e.container;if(!t.disableScrollLock){if(AS(i)){const n=M_(_t(i));a.push({value:i.style.paddingRight,property:"padding-right",el:i}),i.style.paddingRight=`${qh(i)+n}px`;const o=_t(i).querySelectorAll(".mui-fixed");[].forEach.call(o,l=>{a.push({value:l.style.paddingRight,property:"padding-right",el:l}),l.style.paddingRight=`${qh(l)+n}px`})}let r;if(i.parentNode instanceof DocumentFragment)r=_t(i).body;else{const n=i.parentElement,o=ha(i);r=(n==null?void 0:n.nodeName)==="HTML"&&o.getComputedStyle(n).overflowY==="scroll"?n:i}a.push({value:r.style.overflow,property:"overflow",el:r},{value:r.style.overflowX,property:"overflow-x",el:r},{value:r.style.overflowY,property:"overflow-y",el:r}),r.style.overflow="hidden"}return()=>{a.forEach(({value:r,el:n,property:o})=>{r?n.style.setProperty(o,r):n.style.removeProperty(o)})}}function NS(e){const t=[];return[].forEach.call(e.children,a=>{a.getAttribute("aria-hidden")==="true"&&t.push(a)}),t}class jS{constructor(){this.containers=void 0,this.modals=void 0,this.modals=[],this.containers=[]}add(t,a){let i=this.modals.indexOf(t);if(i!==-1)return i;i=this.modals.length,this.modals.push(t),t.modalRef&&wr(t.modalRef,!1);const s=NS(a);Dh(a,t.mount,t.modalRef,s,!0);const r=Bd(this.containers,n=>n.container===a);return r!==-1?(this.containers[r].modals.push(t),i):(this.containers.push({modals:[t],container:a,restore:null,hiddenSiblings:s}),i)}mount(t,a){const i=Bd(this.containers,r=>r.modals.indexOf(t)!==-1),s=this.containers[i];s.restore||(s.restore=IS(s,a))}remove(t,a=!0){const i=this.modals.indexOf(t);if(i===-1)return i;const s=Bd(this.containers,n=>n.modals.indexOf(t)!==-1),r=this.containers[s];if(r.modals.splice(r.modals.indexOf(t),1),this.modals.splice(i,1),r.modals.length===0)r.restore&&r.restore(),t.modalRef&&wr(t.modalRef,a),Dh(r.container,t.mount,t.modalRef,r.hiddenSiblings,!1),this.containers.splice(s,1);else{const n=r.modals[r.modals.length-1];n.modalRef&&wr(n.modalRef,!1)}return i}isTopModal(t){return this.modals.length>0&&this.modals[this.modals.length-1]===t}}const ES=["input","select","textarea","a[href]","button","[tabindex]","audio[controls]","video[controls]",'[contenteditable]:not([contenteditable="false"])'].join(",");function FS(e){const t=parseInt(e.getAttribute("tabindex")||"",10);return Number.isNaN(t)?e.contentEditable==="true"||(e.nodeName==="AUDIO"||e.nodeName==="VIDEO"||e.nodeName==="DETAILS")&&e.getAttribute("tabindex")===null?0:e.tabIndex:t}function BS(e){if(e.tagName!=="INPUT"||e.type!=="radio"||!e.name)return!1;const t=i=>e.ownerDocument.querySelector(`input[type="radio"]${i}`);let a=t(`[name="${e.name}"]:checked`);return a||(a=t(`[name="${e.name}"]`)),a!==e}function OS(e){return!(e.disabled||e.tagName==="INPUT"&&e.type==="hidden"||BS(e))}function MS(e){const t=[],a=[];return Array.from(e.querySelectorAll(ES)).forEach((i,s)=>{const r=FS(i);r===-1||!OS(i)||(r===0?t.push(i):a.push({documentOrder:s,tabIndex:r,node:i}))}),a.sort((i,s)=>i.tabIndex===s.tabIndex?i.documentOrder-s.documentOrder:i.tabIndex-s.tabIndex).map(i=>i.node).concat(t)}function LS(){return!0}function WS(e){const{children:t,disableAutoFocus:a=!1,disableEnforceFocus:i=!1,disableRestoreFocus:s=!1,getTabbable:r=MS,isEnabled:n=LS,open:o}=e,l=b.useRef(!1),d=b.useRef(null),c=b.useRef(null),p=b.useRef(null),u=b.useRef(null),v=b.useRef(!1),f=b.useRef(null),h=Ke(t.ref,f),T=b.useRef(null);b.useEffect(()=>{!o||!f.current||(v.current=!a)},[a,o]),b.useEffect(()=>{if(!o||!f.current)return;const _=_t(f.current);return f.current.contains(_.activeElement)||(f.current.hasAttribute("tabIndex")||f.current.setAttribute("tabIndex","-1"),v.current&&f.current.focus()),()=>{s||(p.current&&p.current.focus&&(l.current=!0,p.current.focus()),p.current=null)}},[o]),b.useEffect(()=>{if(!o||!f.current)return;const _=_t(f.current),w=k=>{T.current=k,!(i||!n()||k.key!=="Tab")&&_.activeElement===f.current&&k.shiftKey&&(l.current=!0,c.current&&c.current.focus())},D=()=>{const k=f.current;if(k===null)return;if(!_.hasFocus()||!n()||l.current){l.current=!1;return}if(k.contains(_.activeElement)||i&&_.activeElement!==d.current&&_.activeElement!==c.current)return;if(_.activeElement!==u.current)u.current=null;else if(u.current!==null)return;if(!v.current)return;let S=[];if((_.activeElement===d.current||_.activeElement===c.current)&&(S=r(f.current)),S.length>0){var A,P;const j=!!((A=T.current)!=null&&A.shiftKey&&((P=T.current)==null?void 0:P.key)==="Tab"),E=S[0],F=S[S.length-1];typeof E!="string"&&typeof F!="string"&&(j?F.focus():E.focus())}else k.focus()};_.addEventListener("focusin",D),_.addEventListener("keydown",w,!0);const x=setInterval(()=>{_.activeElement&&_.activeElement.tagName==="BODY"&&D()},50);return()=>{clearInterval(x),_.removeEventListener("focusin",D),_.removeEventListener("keydown",w,!0)}},[a,i,s,n,o,r]);const y=_=>{p.current===null&&(p.current=_.relatedTarget),v.current=!0,u.current=_.target;const w=t.props.onFocus;w&&w(_)},g=_=>{p.current===null&&(p.current=_.relatedTarget),v.current=!0};return m.jsxs(b.Fragment,{children:[m.jsx("div",{tabIndex:o?0:-1,onFocus:g,ref:d,"data-testid":"sentinelStart"}),b.cloneElement(t,{ref:h,onFocus:y}),m.jsx("div",{tabIndex:o?0:-1,onFocus:g,ref:c,"data-testid":"sentinelEnd"})]})}function zS(e){return typeof e=="function"?e():e}const $S=b.forwardRef(function(t,a){const{children:i,container:s,disablePortal:r=!1}=t,[n,o]=b.useState(null),l=Ke(b.isValidElement(i)?i.ref:null,a);if(Ia(()=>{r||o(zS(s)||document.body)},[s,r]),Ia(()=>{if(n&&!r)return Io(a,n),()=>{Io(a,null)}},[a,n,r]),r){if(b.isValidElement(i)){const d={ref:l};return b.cloneElement(i,d)}return m.jsx(b.Fragment,{children:i})}return m.jsx(b.Fragment,{children:n&&cp.createPortal(i,n)})}),US=$S,KS=["addEndListener","appear","children","easing","in","onEnter","onEntered","onEntering","onExit","onExited","onExiting","style","timeout","TransitionComponent"],VS={entering:{opacity:1},entered:{opacity:1}},HS=b.forwardRef(function(t,a){const i=on(),s={enter:i.transitions.duration.enteringScreen,exit:i.transitions.duration.leavingScreen},{addEndListener:r,appear:n=!0,children:o,easing:l,in:d,onEnter:c,onEntered:p,onEntering:u,onExit:v,onExited:f,onExiting:h,style:T,timeout:y=s,TransitionComponent:g=Wp}=t,_=U(t,KS),w=b.useRef(null),D=Ke(w,o.ref,a),x=I=>B=>{if(I){const O=w.current;B===void 0?I(O):I(O,B)}},k=x(u),S=x((I,B)=>{Up(I);const O=Ps({style:T,timeout:y,easing:l},{mode:"enter"});I.style.webkitTransition=i.transitions.create("opacity",O),I.style.transition=i.transitions.create("opacity",O),c&&c(I,B)}),A=x(p),P=x(h),j=x(I=>{const B=Ps({style:T,timeout:y,easing:l},{mode:"exit"});I.style.webkitTransition=i.transitions.create("opacity",B),I.style.transition=i.transitions.create("opacity",B),v&&v(I)}),E=x(f),F=I=>{r&&r(w.current,I)};return m.jsx(g,q({appear:n,in:d,nodeRef:w,onEnter:S,onEntered:A,onEntering:k,onExit:j,onExited:E,onExiting:P,addEndListener:F,timeout:y},_,{children:(I,B)=>b.cloneElement(o,q({style:q({opacity:0,visibility:I==="exited"&&!d?"hidden":void 0},VS[I],T,o.props.style),ref:D},B))}))}),Qv=HS;function GS(e){return pe("MuiBackdrop",e)}re("MuiBackdrop",["root","invisible"]);const XS=["children","className","component","components","componentsProps","invisible","open","slotProps","slots","TransitionComponent","transitionDuration"],JS=e=>{const{classes:t,invisible:a}=e;return fe({root:["root",a&&"invisible"]},GS,t)},QS=z("div",{name:"MuiBackdrop",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,a.invisible&&t.invisible]}})(({ownerState:e})=>q({position:"fixed",display:"flex",alignItems:"center",justifyContent:"center",right:0,bottom:0,top:0,left:0,backgroundColor:"rgba(0, 0, 0, 0.5)",WebkitTapHighlightColor:"transparent"},e.invisible&&{backgroundColor:"transparent"})),YS=b.forwardRef(function(t,a){var i,s,r;const n=me({props:t,name:"MuiBackdrop"}),{children:o,className:l,component:d="div",components:c={},componentsProps:p={},invisible:u=!1,open:v,slotProps:f={},slots:h={},TransitionComponent:T=Qv,transitionDuration:y}=n,g=U(n,XS),_=q({},n,{component:d,invisible:u}),w=JS(_),D=(i=f.root)!=null?i:p.root;return m.jsx(T,q({in:v,timeout:y},g,{children:m.jsx(QS,q({"aria-hidden":!0},D,{as:(s=(r=h.root)!=null?r:c.Root)!=null?s:d,className:H(w.root,l,D==null?void 0:D.className),ownerState:q({},_,D==null?void 0:D.ownerState),classes:w,ref:a,children:o}))}))}),Yv=YS;function ZS(e){return typeof e=="function"?e():e}function eP(e){return e?e.props.hasOwnProperty("in"):!1}const tP=new jS;function aP(e){const{container:t,disableEscapeKeyDown:a=!1,disableScrollLock:i=!1,manager:s=tP,closeAfterTransition:r=!1,onTransitionEnter:n,onTransitionExited:o,children:l,onClose:d,open:c,rootRef:p}=e,u=b.useRef({}),v=b.useRef(null),f=b.useRef(null),h=Ke(f,p),[T,y]=b.useState(!c),g=eP(l);let _=!0;(e["aria-hidden"]==="false"||e["aria-hidden"]===!1)&&(_=!1);const w=()=>_t(v.current),D=()=>(u.current.modalRef=f.current,u.current.mount=v.current,u.current),x=()=>{s.mount(D(),{disableScrollLock:i}),f.current&&(f.current.scrollTop=0)},k=Si(()=>{const O=ZS(t)||w().body;s.add(D(),O),f.current&&x()}),S=b.useCallback(()=>s.isTopModal(D()),[s]),A=Si(O=>{v.current=O,O&&(c&&S()?x():f.current&&wr(f.current,_))}),P=b.useCallback(()=>{s.remove(D(),_)},[_,s]);b.useEffect(()=>()=>{P()},[P]),b.useEffect(()=>{c?k():(!g||!r)&&P()},[c,P,g,r,k]);const j=O=>W=>{var C;(C=O.onKeyDown)==null||C.call(O,W),!(W.key!=="Escape"||W.which===229||!S())&&(a||(W.stopPropagation(),d&&d(W,"escapeKeyDown")))},E=O=>W=>{var C;(C=O.onClick)==null||C.call(O,W),W.target===W.currentTarget&&d&&d(W,"backdropClick")};return{getRootProps:(O={})=>{const W=z_(e);delete W.onTransitionEnter,delete W.onTransitionExited;const C=q({},W,O);return q({role:"presentation"},C,{onKeyDown:j(C),ref:h})},getBackdropProps:(O={})=>{const W=O;return q({"aria-hidden":!0},W,{onClick:E(W),open:c})},getTransitionProps:()=>{const O=()=>{y(!1),n&&n()},W=()=>{y(!0),o&&o(),r&&P()};return{onEnter:$u(O,l==null?void 0:l.props.onEnter),onExited:$u(W,l==null?void 0:l.props.onExited)}},rootRef:h,portalRef:A,isTopModal:S,exited:T,hasTransition:g}}function iP(e){return pe("MuiModal",e)}re("MuiModal",["root","hidden","backdrop"]);const sP=["BackdropComponent","BackdropProps","classes","className","closeAfterTransition","children","container","component","components","componentsProps","disableAutoFocus","disableEnforceFocus","disableEscapeKeyDown","disablePortal","disableRestoreFocus","disableScrollLock","hideBackdrop","keepMounted","onBackdropClick","onClose","onTransitionEnter","onTransitionExited","open","slotProps","slots","theme"],rP=e=>{const{open:t,exited:a,classes:i}=e;return fe({root:["root",!t&&a&&"hidden"],backdrop:["backdrop"]},iP,i)},nP=z("div",{name:"MuiModal",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,!a.open&&a.exited&&t.hidden]}})(({theme:e,ownerState:t})=>q({position:"fixed",zIndex:(e.vars||e).zIndex.modal,right:0,bottom:0,top:0,left:0},!t.open&&t.exited&&{visibility:"hidden"})),oP=z(Yv,{name:"MuiModal",slot:"Backdrop",overridesResolver:(e,t)=>t.backdrop})({zIndex:-1}),lP=b.forwardRef(function(t,a){var i,s,r,n,o,l;const d=me({name:"MuiModal",props:t}),{BackdropComponent:c=oP,BackdropProps:p,className:u,closeAfterTransition:v=!1,children:f,container:h,component:T,components:y={},componentsProps:g={},disableAutoFocus:_=!1,disableEnforceFocus:w=!1,disableEscapeKeyDown:D=!1,disablePortal:x=!1,disableRestoreFocus:k=!1,disableScrollLock:S=!1,hideBackdrop:A=!1,keepMounted:P=!1,onBackdropClick:j,open:E,slotProps:F,slots:I}=d,B=U(d,sP),O=q({},d,{closeAfterTransition:v,disableAutoFocus:_,disableEnforceFocus:w,disableEscapeKeyDown:D,disablePortal:x,disableRestoreFocus:k,disableScrollLock:S,hideBackdrop:A,keepMounted:P}),{getRootProps:W,getBackdropProps:C,getTransitionProps:N,portalRef:L,isTopModal:te,exited:J,hasTransition:Ne}=aP(q({},O,{rootRef:a})),ee=q({},O,{exited:J}),be=rP(ee),oe={};if(f.props.tabIndex===void 0&&(oe.tabIndex="-1"),Ne){const{onEnter:Z,onExited:ce}=N();oe.onEnter=Z,oe.onExited=ce}const He=(i=(s=I==null?void 0:I.root)!=null?s:y.Root)!=null?i:nP,Nt=(r=(n=I==null?void 0:I.backdrop)!=null?n:y.Backdrop)!=null?r:c,je=(o=F==null?void 0:F.root)!=null?o:g.root,ut=(l=F==null?void 0:F.backdrop)!=null?l:g.backdrop,ne=Cs({elementType:He,externalSlotProps:je,externalForwardedProps:B,getSlotProps:W,additionalProps:{ref:a,as:T},ownerState:ee,className:H(u,je==null?void 0:je.className,be==null?void 0:be.root,!ee.open&&ee.exited&&(be==null?void 0:be.hidden))}),De=Cs({elementType:Nt,externalSlotProps:ut,additionalProps:p,getSlotProps:Z=>C(q({},Z,{onClick:ce=>{j&&j(ce),Z!=null&&Z.onClick&&Z.onClick(ce)}})),className:H(ut==null?void 0:ut.className,p==null?void 0:p.className,be==null?void 0:be.backdrop),ownerState:ee});return!P&&!E&&(!Ne||J)?null:m.jsx(US,{ref:L,container:h,disablePortal:x,children:m.jsxs(He,q({},ne,{children:[!A&&c?m.jsx(Nt,q({},De)):null,m.jsx(WS,{disableEnforceFocus:w,disableAutoFocus:_,disableRestoreFocus:k,isEnabled:te,open:E,children:b.cloneElement(f,oe)})]}))})}),Kp=lP;function dP(e){return pe("MuiPaper",e)}re("MuiPaper",["root","rounded","outlined","elevation","elevation0","elevation1","elevation2","elevation3","elevation4","elevation5","elevation6","elevation7","elevation8","elevation9","elevation10","elevation11","elevation12","elevation13","elevation14","elevation15","elevation16","elevation17","elevation18","elevation19","elevation20","elevation21","elevation22","elevation23","elevation24"]);const uP=["className","component","elevation","square","variant"],cP=e=>{const{square:t,elevation:a,variant:i,classes:s}=e,r={root:["root",i,!t&&"rounded",i==="elevation"&&`elevation${a}`]};return fe(r,dP,s)},pP=z("div",{name:"MuiPaper",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,t[a.variant],!a.square&&t.rounded,a.variant==="elevation"&&t[`elevation${a.elevation}`]]}})(({theme:e,ownerState:t})=>{var a;return q({backgroundColor:(e.vars||e).palette.background.paper,color:(e.vars||e).palette.text.primary,transition:e.transitions.create("box-shadow")},!t.square&&{borderRadius:e.shape.borderRadius},t.variant==="outlined"&&{border:`1px solid ${(e.vars||e).palette.divider}`},t.variant==="elevation"&&q({boxShadow:(e.vars||e).shadows[t.elevation]},!e.vars&&e.palette.mode==="dark"&&{backgroundImage:`linear-gradient(${xa("#fff",ch(t.elevation))}, ${xa("#fff",ch(t.elevation))})`},e.vars&&{backgroundImage:(a=e.vars.overlays)==null?void 0:a[t.elevation]}))}),fP=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiPaper"}),{className:s,component:r="div",elevation:n=1,square:o=!1,variant:l="elevation"}=i,d=U(i,uP),c=q({},i,{component:r,elevation:n,square:o,variant:l}),p=cP(c);return m.jsx(pP,q({as:r,ownerState:c,className:H(p.root,s),ref:a},d))}),un=fP;function mP(e){return pe("MuiPopover",e)}re("MuiPopover",["root","paper"]);const hP=["onEntering"],gP=["action","anchorEl","anchorOrigin","anchorPosition","anchorReference","children","className","container","elevation","marginThreshold","open","PaperProps","slots","slotProps","transformOrigin","TransitionComponent","transitionDuration","TransitionProps","disableScrollLock"],yP=["slotProps"];function xh(e,t){let a=0;return typeof t=="number"?a=t:t==="center"?a=e.height/2:t==="bottom"&&(a=e.height),a}function Ch(e,t){let a=0;return typeof t=="number"?a=t:t==="center"?a=e.width/2:t==="right"&&(a=e.width),a}function Sh(e){return[e.horizontal,e.vertical].map(t=>typeof t=="number"?`${t}px`:t).join(" ")}function Od(e){return typeof e=="function"?e():e}const _P=e=>{const{classes:t}=e;return fe({root:["root"],paper:["paper"]},mP,t)},vP=z(Kp,{name:"MuiPopover",slot:"Root",overridesResolver:(e,t)=>t.root})({}),Zv=z(un,{name:"MuiPopover",slot:"Paper",overridesResolver:(e,t)=>t.paper})({position:"absolute",overflowY:"auto",overflowX:"hidden",minWidth:16,minHeight:16,maxWidth:"calc(100% - 32px)",maxHeight:"calc(100% - 32px)",outline:0}),bP=b.forwardRef(function(t,a){var i,s,r;const n=me({props:t,name:"MuiPopover"}),{action:o,anchorEl:l,anchorOrigin:d={vertical:"top",horizontal:"left"},anchorPosition:c,anchorReference:p="anchorEl",children:u,className:v,container:f,elevation:h=8,marginThreshold:T=16,open:y,PaperProps:g={},slots:_,slotProps:w,transformOrigin:D={vertical:"top",horizontal:"left"},TransitionComponent:x=PS,transitionDuration:k="auto",TransitionProps:{onEntering:S}={},disableScrollLock:A=!1}=n,P=U(n.TransitionProps,hP),j=U(n,gP),E=(i=w==null?void 0:w.paper)!=null?i:g,F=b.useRef(),I=Ke(F,E.ref),B=q({},n,{anchorOrigin:d,anchorReference:p,elevation:h,marginThreshold:T,externalPaperSlotProps:E,transformOrigin:D,TransitionComponent:x,transitionDuration:k,TransitionProps:P}),O=_P(B),W=b.useCallback(()=>{if(p==="anchorPosition")return c;const Z=Od(l),we=(Z&&Z.nodeType===1?Z:_t(F.current).body).getBoundingClientRect();return{top:we.top+xh(we,d.vertical),left:we.left+Ch(we,d.horizontal)}},[l,d.horizontal,d.vertical,c,p]),C=b.useCallback(Z=>({vertical:xh(Z,D.vertical),horizontal:Ch(Z,D.horizontal)}),[D.horizontal,D.vertical]),N=b.useCallback(Z=>{const ce={width:Z.offsetWidth,height:Z.offsetHeight},we=C(ce);if(p==="none")return{top:null,left:null,transformOrigin:Sh(we)};const Ma=W();let jt=Ma.top-we.vertical,Et=Ma.left-we.horizontal;const Xt=jt+ce.height,Ft=Et+ce.width,xe=ha(Od(l)),_a=xe.innerHeight-T,ct=xe.innerWidth-T;if(T!==null&&jt_a){const Te=Xt-_a;jt-=Te,we.vertical+=Te}if(T!==null&&Etct){const Te=Ft-ct;Et-=Te,we.horizontal+=Te}return{top:`${Math.round(jt)}px`,left:`${Math.round(Et)}px`,transformOrigin:Sh(we)}},[l,p,W,C,T]),[L,te]=b.useState(y),J=b.useCallback(()=>{const Z=F.current;if(!Z)return;const ce=N(Z);ce.top!==null&&(Z.style.top=ce.top),ce.left!==null&&(Z.style.left=ce.left),Z.style.transformOrigin=ce.transformOrigin,te(!0)},[N]);b.useEffect(()=>(A&&window.addEventListener("scroll",J),()=>window.removeEventListener("scroll",J)),[l,A,J]);const Ne=(Z,ce)=>{S&&S(Z,ce),J()},ee=()=>{te(!1)};b.useEffect(()=>{y&&J()}),b.useImperativeHandle(o,()=>y?{updatePosition:()=>{J()}}:null,[y,J]),b.useEffect(()=>{if(!y)return;const Z=hl(()=>{J()}),ce=ha(l);return ce.addEventListener("resize",Z),()=>{Z.clear(),ce.removeEventListener("resize",Z)}},[l,y,J]);let be=k;k==="auto"&&!x.muiSupportAuto&&(be=void 0);const oe=f||(l?_t(Od(l)).body:void 0),He=(s=_==null?void 0:_.root)!=null?s:vP,Nt=(r=_==null?void 0:_.paper)!=null?r:Zv,je=Cs({elementType:Nt,externalSlotProps:q({},E,{style:L?E.style:q({},E.style,{opacity:0})}),additionalProps:{elevation:h,ref:I},ownerState:B,className:H(O.paper,E==null?void 0:E.className)}),ut=Cs({elementType:He,externalSlotProps:(w==null?void 0:w.root)||{},externalForwardedProps:j,additionalProps:{ref:a,slotProps:{backdrop:{invisible:!0}},container:oe,open:y},ownerState:B,className:H(O.root,v)}),{slotProps:ne}=ut,De=U(ut,yP);return m.jsx(He,q({},De,!No(He)&&{slotProps:ne,disableScrollLock:A},{children:m.jsx(x,q({appear:!0,in:y,onEntering:Ne,onExited:ee,timeout:be},P,{children:m.jsx(Nt,q({},je,{children:u}))}))}))}),wP=bP;function TP(e){return pe("MuiMenu",e)}re("MuiMenu",["root","paper","list"]);const kP=["onEntering"],qP=["autoFocus","children","className","disableAutoFocusItem","MenuListProps","onClose","open","PaperProps","PopoverClasses","transitionDuration","TransitionProps","variant","slots","slotProps"],DP={vertical:"top",horizontal:"right"},xP={vertical:"top",horizontal:"left"},CP=e=>{const{classes:t}=e;return fe({root:["root"],paper:["paper"],list:["list"]},TP,t)},SP=z(wP,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiMenu",slot:"Root",overridesResolver:(e,t)=>t.root})({}),PP=z(Zv,{name:"MuiMenu",slot:"Paper",overridesResolver:(e,t)=>t.paper})({maxHeight:"calc(100% - 96px)",WebkitOverflowScrolling:"touch"}),AP=z(xS,{name:"MuiMenu",slot:"List",overridesResolver:(e,t)=>t.list})({outline:0}),RP=b.forwardRef(function(t,a){var i,s;const r=me({props:t,name:"MuiMenu"}),{autoFocus:n=!0,children:o,className:l,disableAutoFocusItem:d=!1,MenuListProps:c={},onClose:p,open:u,PaperProps:v={},PopoverClasses:f,transitionDuration:h="auto",TransitionProps:{onEntering:T}={},variant:y="selectedMenu",slots:g={},slotProps:_={}}=r,w=U(r.TransitionProps,kP),D=U(r,qP),x=Uv(),k=q({},r,{autoFocus:n,disableAutoFocusItem:d,MenuListProps:c,onEntering:T,PaperProps:v,transitionDuration:h,TransitionProps:w,variant:y}),S=CP(k),A=n&&!d&&u,P=b.useRef(null),j=(C,N)=>{P.current&&P.current.adjustStyleForScrollbar(C,{direction:x?"rtl":"ltr"}),T&&T(C,N)},E=C=>{C.key==="Tab"&&(C.preventDefault(),p&&p(C,"tabKeyDown"))};let F=-1;b.Children.map(o,(C,N)=>{b.isValidElement(C)&&(C.props.disabled||(y==="selectedMenu"&&C.props.selected||F===-1)&&(F=N))});const I=(i=g.paper)!=null?i:PP,B=(s=_.paper)!=null?s:v,O=Cs({elementType:g.root,externalSlotProps:_.root,ownerState:k,className:[S.root,l]}),W=Cs({elementType:I,externalSlotProps:B,ownerState:k,className:S.paper});return m.jsx(SP,q({onClose:p,anchorOrigin:{vertical:"bottom",horizontal:x?"right":"left"},transformOrigin:x?DP:xP,slots:{paper:I,root:g.root},slotProps:{root:O,paper:W},open:u,ref:a,transitionDuration:h,TransitionProps:q({onEntering:j},w),ownerState:k},D,{classes:f,children:m.jsx(AP,q({onKeyDown:E,actions:P,autoFocus:n&&(F===-1||d),autoFocusItem:A,variant:y},c,{className:H(S.list,c.className),children:o}))}))}),IP=RP;function NP(e){return pe("MuiNativeSelect",e)}const jP=re("MuiNativeSelect",["root","select","multiple","filled","outlined","standard","disabled","icon","iconOpen","iconFilled","iconOutlined","iconStandard","nativeInput","error"]),Vp=jP,EP=["className","disabled","error","IconComponent","inputRef","variant"],FP=e=>{const{classes:t,variant:a,disabled:i,multiple:s,open:r,error:n}=e,o={select:["select",a,i&&"disabled",s&&"multiple",n&&"error"],icon:["icon",`icon${X(a)}`,r&&"iconOpen",i&&"disabled"]};return fe(o,NP,t)},eb=({ownerState:e,theme:t})=>q({MozAppearance:"none",WebkitAppearance:"none",userSelect:"none",borderRadius:0,cursor:"pointer","&:focus":q({},t.vars?{backgroundColor:`rgba(${t.vars.palette.common.onBackgroundChannel} / 0.05)`}:{backgroundColor:t.palette.mode==="light"?"rgba(0, 0, 0, 0.05)":"rgba(255, 255, 255, 0.05)"},{borderRadius:0}),"&::-ms-expand":{display:"none"},[`&.${Vp.disabled}`]:{cursor:"default"},"&[multiple]":{height:"auto"},"&:not([multiple]) option, &:not([multiple]) optgroup":{backgroundColor:(t.vars||t).palette.background.paper},"&&&":{paddingRight:24,minWidth:16}},e.variant==="filled"&&{"&&&":{paddingRight:32}},e.variant==="outlined"&&{borderRadius:(t.vars||t).shape.borderRadius,"&:focus":{borderRadius:(t.vars||t).shape.borderRadius},"&&&":{paddingRight:32}}),BP=z("select",{name:"MuiNativeSelect",slot:"Select",shouldForwardProp:It,overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.select,t[a.variant],a.error&&t.error,{[`&.${Vp.multiple}`]:t.multiple}]}})(eb),tb=({ownerState:e,theme:t})=>q({position:"absolute",right:0,top:"calc(50% - .5em)",pointerEvents:"none",color:(t.vars||t).palette.action.active,[`&.${Vp.disabled}`]:{color:(t.vars||t).palette.action.disabled}},e.open&&{transform:"rotate(180deg)"},e.variant==="filled"&&{right:7},e.variant==="outlined"&&{right:7}),OP=z("svg",{name:"MuiNativeSelect",slot:"Icon",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.icon,a.variant&&t[`icon${X(a.variant)}`],a.open&&t.iconOpen]}})(tb),MP=b.forwardRef(function(t,a){const{className:i,disabled:s,error:r,IconComponent:n,inputRef:o,variant:l="standard"}=t,d=U(t,EP),c=q({},t,{disabled:s,variant:l,error:r}),p=FP(c);return m.jsxs(b.Fragment,{children:[m.jsx(BP,q({ownerState:c,className:H(p.select,i),disabled:s,ref:o||a},d)),t.multiple?null:m.jsx(OP,{as:n,ownerState:c,className:p.icon})]})}),LP=MP;function WP(e){return pe("MuiSelect",e)}const zP=re("MuiSelect",["root","select","multiple","filled","outlined","standard","disabled","focused","icon","iconOpen","iconFilled","iconOutlined","iconStandard","nativeInput","error"]),er=zP;var Ph;const $P=["aria-describedby","aria-label","autoFocus","autoWidth","children","className","defaultOpen","defaultValue","disabled","displayEmpty","error","IconComponent","inputRef","labelId","MenuProps","multiple","name","onBlur","onChange","onClose","onFocus","onOpen","open","readOnly","renderValue","SelectDisplayProps","tabIndex","type","value","variant"],UP=z("div",{name:"MuiSelect",slot:"Select",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[{[`&.${er.select}`]:t.select},{[`&.${er.select}`]:t[a.variant]},{[`&.${er.error}`]:t.error},{[`&.${er.multiple}`]:t.multiple}]}})(eb,{[`&.${er.select}`]:{height:"auto",minHeight:"1.4375em",textOverflow:"ellipsis",whiteSpace:"nowrap",overflow:"hidden"}}),KP=z("svg",{name:"MuiSelect",slot:"Icon",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.icon,a.variant&&t[`icon${X(a.variant)}`],a.open&&t.iconOpen]}})(tb),VP=z("input",{shouldForwardProp:e=>Sv(e)&&e!=="classes",name:"MuiSelect",slot:"NativeInput",overridesResolver:(e,t)=>t.nativeInput})({bottom:0,left:0,position:"absolute",opacity:0,pointerEvents:"none",width:"100%",boxSizing:"border-box"});function Ah(e,t){return typeof t=="object"&&t!==null?e===t:String(e)===String(t)}function HP(e){return e==null||typeof e=="string"&&!e.trim()}const GP=e=>{const{classes:t,variant:a,disabled:i,multiple:s,open:r,error:n}=e,o={select:["select",a,i&&"disabled",s&&"multiple",n&&"error"],icon:["icon",`icon${X(a)}`,r&&"iconOpen",i&&"disabled"],nativeInput:["nativeInput"]};return fe(o,WP,t)},XP=b.forwardRef(function(t,a){var i;const{"aria-describedby":s,"aria-label":r,autoFocus:n,autoWidth:o,children:l,className:d,defaultOpen:c,defaultValue:p,disabled:u,displayEmpty:v,error:f=!1,IconComponent:h,inputRef:T,labelId:y,MenuProps:g={},multiple:_,name:w,onBlur:D,onChange:x,onClose:k,onFocus:S,onOpen:A,open:P,readOnly:j,renderValue:E,SelectDisplayProps:F={},tabIndex:I,value:B,variant:O="standard"}=t,W=U(t,$P),[C,N]=Uu({controlled:B,default:p,name:"Select"}),[L,te]=Uu({controlled:P,default:c,name:"Select"}),J=b.useRef(null),Ne=b.useRef(null),[ee,be]=b.useState(null),{current:oe}=b.useRef(P!=null),[He,Nt]=b.useState(),je=Ke(a,T),ut=b.useCallback(K=>{Ne.current=K,K&&be(K)},[]),ne=ee==null?void 0:ee.parentNode;b.useImperativeHandle(je,()=>({focus:()=>{Ne.current.focus()},node:J.current,value:C}),[C]),b.useEffect(()=>{c&&L&&ee&&!oe&&(Nt(o?null:ne.clientWidth),Ne.current.focus())},[ee,o]),b.useEffect(()=>{n&&Ne.current.focus()},[n]),b.useEffect(()=>{if(!y)return;const K=_t(Ne.current).getElementById(y);if(K){const ye=()=>{getSelection().isCollapsed&&Ne.current.focus()};return K.addEventListener("click",ye),()=>{K.removeEventListener("click",ye)}}},[y]);const De=(K,ye)=>{K?A&&A(ye):k&&k(ye),oe||(Nt(o?null:ne.clientWidth),te(K))},Z=K=>{K.button===0&&(K.preventDefault(),Ne.current.focus(),De(!0,K))},ce=K=>{De(!1,K)},we=b.Children.toArray(l),Ma=K=>{const ye=we.find(Ge=>Ge.props.value===K.target.value);ye!==void 0&&(N(ye.props.value),x&&x(K,ye))},jt=K=>ye=>{let Ge;if(ye.currentTarget.hasAttribute("tabindex")){if(_){Ge=Array.isArray(C)?C.slice():[];const $i=C.indexOf(K.props.value);$i===-1?Ge.push(K.props.value):Ge.splice($i,1)}else Ge=K.props.value;if(K.props.onClick&&K.props.onClick(ye),C!==Ge&&(N(Ge),x)){const $i=ye.nativeEvent||ye,uf=new $i.constructor($i.type,$i);Object.defineProperty(uf,"target",{writable:!0,value:{value:Ge,name:w}}),x(uf,K)}_||De(!1,ye)}},Et=K=>{j||[" ","ArrowUp","ArrowDown","Enter"].indexOf(K.key)!==-1&&(K.preventDefault(),De(!0,K))},Xt=ee!==null&&L,Ft=K=>{!Xt&&D&&(Object.defineProperty(K,"target",{writable:!0,value:{value:C,name:w}}),D(K))};delete W["aria-invalid"];let xe,_a;const ct=[];let Te=!1;(Mo({value:C})||v)&&(E?xe=E(C):Te=!0);const ra=we.map(K=>{if(!b.isValidElement(K))return null;let ye;if(_){if(!Array.isArray(C))throw new Error(Oi(2));ye=C.some(Ge=>Ah(Ge,K.props.value)),ye&&Te&&ct.push(K.props.children)}else ye=Ah(C,K.props.value),ye&&Te&&(_a=K.props.children);return b.cloneElement(K,{"aria-selected":ye?"true":"false",onClick:jt(K),onKeyUp:Ge=>{Ge.key===" "&&Ge.preventDefault(),K.props.onKeyUp&&K.props.onKeyUp(Ge)},role:"option",selected:ye,value:void 0,"data-value":K.props.value})});Te&&(_?ct.length===0?xe=null:xe=ct.reduce((K,ye,Ge)=>(K.push(ye),Ge{const y=c.current,_=ha(y).getComputedStyle(y);if(_.width==="0px")return{outerHeightStyle:0,overflowing:!1};const w=v.current;w.style.width=_.width,w.value=y.value||t.placeholder||"x",w.value.slice(-1)===` +`&&(w.value+=" ");const D=_.boxSizing,x=In(_.paddingBottom)+In(_.paddingTop),k=In(_.borderBottomWidth)+In(_.borderTopWidth),S=w.scrollHeight;w.value="x";const A=w.scrollHeight;let P=S;r&&(P=Math.max(Number(r)*A,P)),s&&(P=Math.min(Number(s)*A,P)),P=Math.max(P,A);const j=P+(D==="border-box"?x+k:0),E=Math.abs(P-S)<=1;return{outerHeightStyle:j,overflowing:E}},[s,r,t.placeholder]),h=b.useCallback(()=>{const y=f();if(eA(y))return;const g=y.outerHeightStyle,_=c.current;u.current!==g&&(u.current=g,_.style.height=`${g}px`),_.style.overflow=y.overflowing?"hidden":""},[f]);Ia(()=>{const y=()=>{h()};let g;const _=hl(y),w=c.current,D=ha(w);D.addEventListener("resize",_);let x;return typeof ResizeObserver<"u"&&(x=new ResizeObserver(y),x.observe(w)),()=>{_.clear(),cancelAnimationFrame(g),D.removeEventListener("resize",_),x&&x.disconnect()}},[f,h]),Ia(()=>{h()});const T=y=>{d||h(),i&&i(y)};return m.jsxs(b.Fragment,{children:[m.jsx("textarea",q({value:o,onChange:T,ref:p,rows:r,style:n},l)),m.jsx("textarea",{"aria-hidden":!0,className:t.className,readOnly:!0,ref:v,tabIndex:-1,style:q({},ZP.shadow,n,{paddingTop:0,paddingBottom:0})})]})}),aA=tA;function iA(e){return m.jsx(KC,q({},e,{defaultTheme:Ep,themeId:Ll}))}function sA(e){return pe("MuiInputBase",e)}const rA=re("MuiInputBase",["root","formControl","focused","disabled","adornedStart","adornedEnd","error","sizeSmall","multiline","colorSecondary","fullWidth","hiddenLabel","readOnly","input","inputSizeSmall","inputMultiline","inputTypeSearch","inputAdornedStart","inputAdornedEnd","inputHiddenLabel"]),As=rA,nA=["aria-describedby","autoComplete","autoFocus","className","color","components","componentsProps","defaultValue","disabled","disableInjectingGlobalStyles","endAdornment","error","fullWidth","id","inputComponent","inputProps","inputRef","margin","maxRows","minRows","multiline","name","onBlur","onChange","onClick","onFocus","onKeyDown","onKeyUp","placeholder","readOnly","renderSuffix","rows","size","slotProps","slots","startAdornment","type","value"],zl=(e,t)=>{const{ownerState:a}=e;return[t.root,a.formControl&&t.formControl,a.startAdornment&&t.adornedStart,a.endAdornment&&t.adornedEnd,a.error&&t.error,a.size==="small"&&t.sizeSmall,a.multiline&&t.multiline,a.color&&t[`color${X(a.color)}`],a.fullWidth&&t.fullWidth,a.hiddenLabel&&t.hiddenLabel]},$l=(e,t)=>{const{ownerState:a}=e;return[t.input,a.size==="small"&&t.inputSizeSmall,a.multiline&&t.inputMultiline,a.type==="search"&&t.inputTypeSearch,a.startAdornment&&t.inputAdornedStart,a.endAdornment&&t.inputAdornedEnd,a.hiddenLabel&&t.inputHiddenLabel]},oA=e=>{const{classes:t,color:a,disabled:i,error:s,endAdornment:r,focused:n,formControl:o,fullWidth:l,hiddenLabel:d,multiline:c,readOnly:p,size:u,startAdornment:v,type:f}=e,h={root:["root",`color${X(a)}`,i&&"disabled",s&&"error",l&&"fullWidth",n&&"focused",o&&"formControl",u&&u!=="medium"&&`size${X(u)}`,c&&"multiline",v&&"adornedStart",r&&"adornedEnd",d&&"hiddenLabel",p&&"readOnly"],input:["input",i&&"disabled",f==="search"&&"inputTypeSearch",c&&"inputMultiline",u==="small"&&"inputSizeSmall",d&&"inputHiddenLabel",v&&"inputAdornedStart",r&&"inputAdornedEnd",p&&"readOnly"]};return fe(h,sA,t)},Ul=z("div",{name:"MuiInputBase",slot:"Root",overridesResolver:zl})(({theme:e,ownerState:t})=>q({},e.typography.body1,{color:(e.vars||e).palette.text.primary,lineHeight:"1.4375em",boxSizing:"border-box",position:"relative",cursor:"text",display:"inline-flex",alignItems:"center",[`&.${As.disabled}`]:{color:(e.vars||e).palette.text.disabled,cursor:"default"}},t.multiline&&q({padding:"4px 0 5px"},t.size==="small"&&{paddingTop:1}),t.fullWidth&&{width:"100%"})),Kl=z("input",{name:"MuiInputBase",slot:"Input",overridesResolver:$l})(({theme:e,ownerState:t})=>{const a=e.palette.mode==="light",i=q({color:"currentColor"},e.vars?{opacity:e.vars.opacity.inputPlaceholder}:{opacity:a?.42:.5},{transition:e.transitions.create("opacity",{duration:e.transitions.duration.shorter})}),s={opacity:"0 !important"},r=e.vars?{opacity:e.vars.opacity.inputPlaceholder}:{opacity:a?.42:.5};return q({font:"inherit",letterSpacing:"inherit",color:"currentColor",padding:"4px 0 5px",border:0,boxSizing:"content-box",background:"none",height:"1.4375em",margin:0,WebkitTapHighlightColor:"transparent",display:"block",minWidth:0,width:"100%",animationName:"mui-auto-fill-cancel",animationDuration:"10ms","&::-webkit-input-placeholder":i,"&::-moz-placeholder":i,"&:-ms-input-placeholder":i,"&::-ms-input-placeholder":i,"&:focus":{outline:0},"&:invalid":{boxShadow:"none"},"&::-webkit-search-decoration":{WebkitAppearance:"none"},[`label[data-shrink=false] + .${As.formControl} &`]:{"&::-webkit-input-placeholder":s,"&::-moz-placeholder":s,"&:-ms-input-placeholder":s,"&::-ms-input-placeholder":s,"&:focus::-webkit-input-placeholder":r,"&:focus::-moz-placeholder":r,"&:focus:-ms-input-placeholder":r,"&:focus::-ms-input-placeholder":r},[`&.${As.disabled}`]:{opacity:1,WebkitTextFillColor:(e.vars||e).palette.text.disabled},"&:-webkit-autofill":{animationDuration:"5000s",animationName:"mui-auto-fill"}},t.size==="small"&&{paddingTop:1},t.multiline&&{height:"auto",resize:"none",padding:0,paddingTop:0},t.type==="search"&&{MozAppearance:"textfield"})}),lA=m.jsx(iA,{styles:{"@keyframes mui-auto-fill":{from:{display:"block"}},"@keyframes mui-auto-fill-cancel":{from:{display:"block"}}}}),dA=b.forwardRef(function(t,a){var i;const s=me({props:t,name:"MuiInputBase"}),{"aria-describedby":r,autoComplete:n,autoFocus:o,className:l,components:d={},componentsProps:c={},defaultValue:p,disabled:u,disableInjectingGlobalStyles:v,endAdornment:f,fullWidth:h=!1,id:T,inputComponent:y="input",inputProps:g={},inputRef:_,maxRows:w,minRows:D,multiline:x=!1,name:k,onBlur:S,onChange:A,onClick:P,onFocus:j,onKeyDown:E,onKeyUp:F,placeholder:I,readOnly:B,renderSuffix:O,rows:W,slotProps:C={},slots:N={},startAdornment:L,type:te="text",value:J}=s,Ne=U(s,nA),ee=g.value!=null?g.value:J,{current:be}=b.useRef(ee!=null),oe=b.useRef(),He=b.useCallback(he=>{},[]),Nt=Ke(oe,_,g.ref,He),[je,ut]=b.useState(!1),ne=dn(),De=ln({props:s,muiFormControl:ne,states:["color","disabled","error","hiddenLabel","size","required","filled"]});De.focused=ne?ne.focused:je,b.useEffect(()=>{!ne&&u&&je&&(ut(!1),S&&S())},[ne,u,je,S]);const Z=ne&&ne.onFilled,ce=ne&&ne.onEmpty,we=b.useCallback(he=>{Mo(he)?Z&&Z():ce&&ce()},[Z,ce]);Ia(()=>{be&&we({value:ee})},[ee,we,be]);const Ma=he=>{if(De.disabled){he.stopPropagation();return}j&&j(he),g.onFocus&&g.onFocus(he),ne&&ne.onFocus?ne.onFocus(he):ut(!0)},jt=he=>{S&&S(he),g.onBlur&&g.onBlur(he),ne&&ne.onBlur?ne.onBlur(he):ut(!1)},Et=(he,...V)=>{if(!be){const ba=he.target||oe.current;if(ba==null)throw new Error(Oi(1));we({value:ba.value})}g.onChange&&g.onChange(he,...V),A&&A(he,...V)};b.useEffect(()=>{we(oe.current)},[]);const Xt=he=>{oe.current&&he.currentTarget===he.target&&oe.current.focus(),P&&P(he)};let Ft=y,xe=g;x&&Ft==="input"&&(W?xe=q({type:void 0,minRows:W,maxRows:W},xe):xe=q({type:void 0,maxRows:w,minRows:D},xe),Ft=aA);const _a=he=>{we(he.animationName==="mui-auto-fill-cancel"?oe.current:{value:"x"})};b.useEffect(()=>{ne&&ne.setAdornedStart(!!L)},[ne,L]);const ct=q({},s,{color:De.color||"primary",disabled:De.disabled,endAdornment:f,error:De.error,focused:De.focused,formControl:ne,fullWidth:h,hiddenLabel:De.hiddenLabel,multiline:x,size:De.size,startAdornment:L,type:te}),Te=oA(ct),ra=N.root||d.Root||Ul,La=C.root||c.root||{},va=N.input||d.Input||Kl;return xe=q({},xe,(i=C.input)!=null?i:c.input),m.jsxs(b.Fragment,{children:[!v&&lA,m.jsxs(ra,q({},La,!No(ra)&&{ownerState:q({},ct,La.ownerState)},{ref:a,onClick:Xt},Ne,{className:H(Te.root,La.className,l,B&&"MuiInputBase-readOnly"),children:[L,m.jsx(Lp.Provider,{value:null,children:m.jsx(va,q({ownerState:ct,"aria-invalid":De.error,"aria-describedby":r,autoComplete:n,autoFocus:o,defaultValue:p,disabled:De.disabled,id:T,onAnimationStart:_a,name:k,placeholder:I,readOnly:B,required:De.required,rows:W,value:ee,onKeyDown:E,onKeyUp:F,type:te},xe,!No(va)&&{as:Ft,ownerState:q({},ct,xe.ownerState)},{ref:Nt,className:H(Te.input,xe.className,B&&"MuiInputBase-readOnly"),onBlur:jt,onChange:Et,onFocus:Ma}))}),f,O?O(q({},De,{startAdornment:L})):null]}))]})}),Hp=dA;function uA(e){return pe("MuiInput",e)}const cA=q({},As,re("MuiInput",["root","underline","input"])),tr=cA,pA=["disableUnderline","components","componentsProps","fullWidth","inputComponent","multiline","slotProps","slots","type"],fA=e=>{const{classes:t,disableUnderline:a}=e,s=fe({root:["root",!a&&"underline"],input:["input"]},uA,t);return q({},t,s)},mA=z(Ul,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiInput",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[...zl(e,t),!a.disableUnderline&&t.underline]}})(({theme:e,ownerState:t})=>{let i=e.palette.mode==="light"?"rgba(0, 0, 0, 0.42)":"rgba(255, 255, 255, 0.7)";return e.vars&&(i=`rgba(${e.vars.palette.common.onBackgroundChannel} / ${e.vars.opacity.inputUnderline})`),q({position:"relative"},t.formControl&&{"label + &":{marginTop:16}},!t.disableUnderline&&{"&::after":{borderBottom:`2px solid ${(e.vars||e).palette[t.color].main}`,left:0,bottom:0,content:'""',position:"absolute",right:0,transform:"scaleX(0)",transition:e.transitions.create("transform",{duration:e.transitions.duration.shorter,easing:e.transitions.easing.easeOut}),pointerEvents:"none"},[`&.${tr.focused}:after`]:{transform:"scaleX(1) translateX(0)"},[`&.${tr.error}`]:{"&::before, &::after":{borderBottomColor:(e.vars||e).palette.error.main}},"&::before":{borderBottom:`1px solid ${i}`,left:0,bottom:0,content:'"\\00a0"',position:"absolute",right:0,transition:e.transitions.create("border-bottom-color",{duration:e.transitions.duration.shorter}),pointerEvents:"none"},[`&:hover:not(.${tr.disabled}, .${tr.error}):before`]:{borderBottom:`2px solid ${(e.vars||e).palette.text.primary}`,"@media (hover: none)":{borderBottom:`1px solid ${i}`}},[`&.${tr.disabled}:before`]:{borderBottomStyle:"dotted"}})}),hA=z(Kl,{name:"MuiInput",slot:"Input",overridesResolver:$l})({}),ab=b.forwardRef(function(t,a){var i,s,r,n;const o=me({props:t,name:"MuiInput"}),{disableUnderline:l,components:d={},componentsProps:c,fullWidth:p=!1,inputComponent:u="input",multiline:v=!1,slotProps:f,slots:h={},type:T="text"}=o,y=U(o,pA),g=fA(o),w={root:{ownerState:{disableUnderline:l}}},D=f??c?xt(f??c,w):w,x=(i=(s=h.root)!=null?s:d.Root)!=null?i:mA,k=(r=(n=h.input)!=null?n:d.Input)!=null?r:hA;return m.jsx(Hp,q({slots:{root:x,input:k},slotProps:D,fullWidth:p,inputComponent:u,multiline:v,ref:a,type:T},y,{classes:g}))});ab.muiName="Input";const gA=ab;function yA(e){return pe("MuiFilledInput",e)}const _A=q({},As,re("MuiFilledInput",["root","underline","input"])),yi=_A,vA=["disableUnderline","components","componentsProps","fullWidth","hiddenLabel","inputComponent","multiline","slotProps","slots","type"],bA=e=>{const{classes:t,disableUnderline:a}=e,s=fe({root:["root",!a&&"underline"],input:["input"]},yA,t);return q({},t,s)},wA=z(Ul,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiFilledInput",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[...zl(e,t),!a.disableUnderline&&t.underline]}})(({theme:e,ownerState:t})=>{var a;const i=e.palette.mode==="light",s=i?"rgba(0, 0, 0, 0.42)":"rgba(255, 255, 255, 0.7)",r=i?"rgba(0, 0, 0, 0.06)":"rgba(255, 255, 255, 0.09)",n=i?"rgba(0, 0, 0, 0.09)":"rgba(255, 255, 255, 0.13)",o=i?"rgba(0, 0, 0, 0.12)":"rgba(255, 255, 255, 0.12)";return q({position:"relative",backgroundColor:e.vars?e.vars.palette.FilledInput.bg:r,borderTopLeftRadius:(e.vars||e).shape.borderRadius,borderTopRightRadius:(e.vars||e).shape.borderRadius,transition:e.transitions.create("background-color",{duration:e.transitions.duration.shorter,easing:e.transitions.easing.easeOut}),"&:hover":{backgroundColor:e.vars?e.vars.palette.FilledInput.hoverBg:n,"@media (hover: none)":{backgroundColor:e.vars?e.vars.palette.FilledInput.bg:r}},[`&.${yi.focused}`]:{backgroundColor:e.vars?e.vars.palette.FilledInput.bg:r},[`&.${yi.disabled}`]:{backgroundColor:e.vars?e.vars.palette.FilledInput.disabledBg:o}},!t.disableUnderline&&{"&::after":{borderBottom:`2px solid ${(a=(e.vars||e).palette[t.color||"primary"])==null?void 0:a.main}`,left:0,bottom:0,content:'""',position:"absolute",right:0,transform:"scaleX(0)",transition:e.transitions.create("transform",{duration:e.transitions.duration.shorter,easing:e.transitions.easing.easeOut}),pointerEvents:"none"},[`&.${yi.focused}:after`]:{transform:"scaleX(1) translateX(0)"},[`&.${yi.error}`]:{"&::before, &::after":{borderBottomColor:(e.vars||e).palette.error.main}},"&::before":{borderBottom:`1px solid ${e.vars?`rgba(${e.vars.palette.common.onBackgroundChannel} / ${e.vars.opacity.inputUnderline})`:s}`,left:0,bottom:0,content:'"\\00a0"',position:"absolute",right:0,transition:e.transitions.create("border-bottom-color",{duration:e.transitions.duration.shorter}),pointerEvents:"none"},[`&:hover:not(.${yi.disabled}, .${yi.error}):before`]:{borderBottom:`1px solid ${(e.vars||e).palette.text.primary}`},[`&.${yi.disabled}:before`]:{borderBottomStyle:"dotted"}},t.startAdornment&&{paddingLeft:12},t.endAdornment&&{paddingRight:12},t.multiline&&q({padding:"25px 12px 8px"},t.size==="small"&&{paddingTop:21,paddingBottom:4},t.hiddenLabel&&{paddingTop:16,paddingBottom:17},t.hiddenLabel&&t.size==="small"&&{paddingTop:8,paddingBottom:9}))}),TA=z(Kl,{name:"MuiFilledInput",slot:"Input",overridesResolver:$l})(({theme:e,ownerState:t})=>q({paddingTop:25,paddingRight:12,paddingBottom:8,paddingLeft:12},!e.vars&&{"&:-webkit-autofill":{WebkitBoxShadow:e.palette.mode==="light"?null:"0 0 0 100px #266798 inset",WebkitTextFillColor:e.palette.mode==="light"?null:"#fff",caretColor:e.palette.mode==="light"?null:"#fff",borderTopLeftRadius:"inherit",borderTopRightRadius:"inherit"}},e.vars&&{"&:-webkit-autofill":{borderTopLeftRadius:"inherit",borderTopRightRadius:"inherit"},[e.getColorSchemeSelector("dark")]:{"&:-webkit-autofill":{WebkitBoxShadow:"0 0 0 100px #266798 inset",WebkitTextFillColor:"#fff",caretColor:"#fff"}}},t.size==="small"&&{paddingTop:21,paddingBottom:4},t.hiddenLabel&&{paddingTop:16,paddingBottom:17},t.startAdornment&&{paddingLeft:0},t.endAdornment&&{paddingRight:0},t.hiddenLabel&&t.size==="small"&&{paddingTop:8,paddingBottom:9},t.multiline&&{paddingTop:0,paddingBottom:0,paddingLeft:0,paddingRight:0})),ib=b.forwardRef(function(t,a){var i,s,r,n;const o=me({props:t,name:"MuiFilledInput"}),{components:l={},componentsProps:d,fullWidth:c=!1,inputComponent:p="input",multiline:u=!1,slotProps:v,slots:f={},type:h="text"}=o,T=U(o,vA),y=q({},o,{fullWidth:c,inputComponent:p,multiline:u,type:h}),g=bA(o),_={root:{ownerState:y},input:{ownerState:y}},w=v??d?xt(_,v??d):_,D=(i=(s=f.root)!=null?s:l.Root)!=null?i:wA,x=(r=(n=f.input)!=null?n:l.Input)!=null?r:TA;return m.jsx(Hp,q({slots:{root:D,input:x},componentsProps:w,fullWidth:c,inputComponent:p,multiline:u,ref:a,type:h},T,{classes:g}))});ib.muiName="Input";const kA=ib;var Rh;const qA=["children","classes","className","label","notched"],DA=z("fieldset",{shouldForwardProp:It})({textAlign:"left",position:"absolute",bottom:0,right:0,top:-5,left:0,margin:0,padding:"0 8px",pointerEvents:"none",borderRadius:"inherit",borderStyle:"solid",borderWidth:1,overflow:"hidden",minWidth:"0%"}),xA=z("legend",{shouldForwardProp:It})(({ownerState:e,theme:t})=>q({float:"unset",width:"auto",overflow:"hidden"},!e.withLabel&&{padding:0,lineHeight:"11px",transition:t.transitions.create("width",{duration:150,easing:t.transitions.easing.easeOut})},e.withLabel&&q({display:"block",padding:0,height:11,fontSize:"0.75em",visibility:"hidden",maxWidth:.01,transition:t.transitions.create("max-width",{duration:50,easing:t.transitions.easing.easeOut}),whiteSpace:"nowrap","& > span":{paddingLeft:5,paddingRight:5,display:"inline-block",opacity:0,visibility:"visible"}},e.notched&&{maxWidth:"100%",transition:t.transitions.create("max-width",{duration:100,easing:t.transitions.easing.easeOut,delay:50})})));function CA(e){const{className:t,label:a,notched:i}=e,s=U(e,qA),r=a!=null&&a!=="",n=q({},e,{notched:i,withLabel:r});return m.jsx(DA,q({"aria-hidden":!0,className:t,ownerState:n},s,{children:m.jsx(xA,{ownerState:n,children:r?m.jsx("span",{children:a}):Rh||(Rh=m.jsx("span",{className:"notranslate",children:"​"}))})}))}function SA(e){return pe("MuiOutlinedInput",e)}const PA=q({},As,re("MuiOutlinedInput",["root","notchedOutline","input"])),za=PA,AA=["components","fullWidth","inputComponent","label","multiline","notched","slots","type"],RA=e=>{const{classes:t}=e,i=fe({root:["root"],notchedOutline:["notchedOutline"],input:["input"]},SA,t);return q({},t,i)},IA=z(Ul,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiOutlinedInput",slot:"Root",overridesResolver:zl})(({theme:e,ownerState:t})=>{const a=e.palette.mode==="light"?"rgba(0, 0, 0, 0.23)":"rgba(255, 255, 255, 0.23)";return q({position:"relative",borderRadius:(e.vars||e).shape.borderRadius,[`&:hover .${za.notchedOutline}`]:{borderColor:(e.vars||e).palette.text.primary},"@media (hover: none)":{[`&:hover .${za.notchedOutline}`]:{borderColor:e.vars?`rgba(${e.vars.palette.common.onBackgroundChannel} / 0.23)`:a}},[`&.${za.focused} .${za.notchedOutline}`]:{borderColor:(e.vars||e).palette[t.color].main,borderWidth:2},[`&.${za.error} .${za.notchedOutline}`]:{borderColor:(e.vars||e).palette.error.main},[`&.${za.disabled} .${za.notchedOutline}`]:{borderColor:(e.vars||e).palette.action.disabled}},t.startAdornment&&{paddingLeft:14},t.endAdornment&&{paddingRight:14},t.multiline&&q({padding:"16.5px 14px"},t.size==="small"&&{padding:"8.5px 14px"}))}),NA=z(CA,{name:"MuiOutlinedInput",slot:"NotchedOutline",overridesResolver:(e,t)=>t.notchedOutline})(({theme:e})=>{const t=e.palette.mode==="light"?"rgba(0, 0, 0, 0.23)":"rgba(255, 255, 255, 0.23)";return{borderColor:e.vars?`rgba(${e.vars.palette.common.onBackgroundChannel} / 0.23)`:t}}),jA=z(Kl,{name:"MuiOutlinedInput",slot:"Input",overridesResolver:$l})(({theme:e,ownerState:t})=>q({padding:"16.5px 14px"},!e.vars&&{"&:-webkit-autofill":{WebkitBoxShadow:e.palette.mode==="light"?null:"0 0 0 100px #266798 inset",WebkitTextFillColor:e.palette.mode==="light"?null:"#fff",caretColor:e.palette.mode==="light"?null:"#fff",borderRadius:"inherit"}},e.vars&&{"&:-webkit-autofill":{borderRadius:"inherit"},[e.getColorSchemeSelector("dark")]:{"&:-webkit-autofill":{WebkitBoxShadow:"0 0 0 100px #266798 inset",WebkitTextFillColor:"#fff",caretColor:"#fff"}}},t.size==="small"&&{padding:"8.5px 14px"},t.multiline&&{padding:0},t.startAdornment&&{paddingLeft:0},t.endAdornment&&{paddingRight:0})),sb=b.forwardRef(function(t,a){var i,s,r,n,o;const l=me({props:t,name:"MuiOutlinedInput"}),{components:d={},fullWidth:c=!1,inputComponent:p="input",label:u,multiline:v=!1,notched:f,slots:h={},type:T="text"}=l,y=U(l,AA),g=RA(l),_=dn(),w=ln({props:l,muiFormControl:_,states:["color","disabled","error","focused","hiddenLabel","size","required"]}),D=q({},l,{color:w.color||"primary",disabled:w.disabled,error:w.error,focused:w.focused,formControl:_,fullWidth:c,hiddenLabel:w.hiddenLabel,multiline:v,size:w.size,type:T}),x=(i=(s=h.root)!=null?s:d.Root)!=null?i:IA,k=(r=(n=h.input)!=null?n:d.Input)!=null?r:jA;return m.jsx(Hp,q({slots:{root:x,input:k},renderSuffix:S=>m.jsx(NA,{ownerState:D,className:g.notchedOutline,label:u!=null&&u!==""&&w.required?o||(o=m.jsxs(b.Fragment,{children:[u," ","*"]})):u,notched:typeof f<"u"?f:!!(S.startAdornment||S.filled||S.focused)}),fullWidth:c,inputComponent:p,multiline:v,ref:a,type:T},y,{classes:q({},g,{notchedOutline:null})}))});sb.muiName="Input";const EA=sb,FA=["autoWidth","children","classes","className","defaultOpen","displayEmpty","IconComponent","id","input","inputProps","label","labelId","MenuProps","multiple","native","onClose","onOpen","open","renderValue","SelectDisplayProps","variant"],BA=["root"],OA=e=>{const{classes:t}=e;return t},Gp={name:"MuiSelect",overridesResolver:(e,t)=>t.root,shouldForwardProp:e=>It(e)&&e!=="variant",slot:"Root"},MA=z(gA,Gp)(""),LA=z(EA,Gp)(""),WA=z(kA,Gp)(""),rb=b.forwardRef(function(t,a){const i=me({name:"MuiSelect",props:t}),{autoWidth:s=!1,children:r,classes:n={},className:o,defaultOpen:l=!1,displayEmpty:d=!1,IconComponent:c=QP,id:p,input:u,inputProps:v,label:f,labelId:h,MenuProps:T,multiple:y=!1,native:g=!1,onClose:_,onOpen:w,open:D,renderValue:x,SelectDisplayProps:k,variant:S="outlined"}=i,A=U(i,FA),P=g?LP:JP,j=dn(),E=ln({props:i,muiFormControl:j,states:["variant","error"]}),F=E.variant||S,I=q({},i,{variant:F,classes:n}),B=OA(I),O=U(B,BA),W=u||{standard:m.jsx(MA,{ownerState:I}),outlined:m.jsx(LA,{label:f,ownerState:I}),filled:m.jsx(WA,{ownerState:I})}[F],C=Ke(a,W.ref);return m.jsx(b.Fragment,{children:b.cloneElement(W,q({inputComponent:P,inputProps:q({children:r,error:E.error,IconComponent:c,variant:F,type:void 0,multiple:y},g?{id:p}:{autoWidth:s,defaultOpen:l,displayEmpty:d,labelId:h,MenuProps:T,onClose:_,onOpen:w,open:D,renderValue:x,SelectDisplayProps:q({id:p},k)},v,{classes:v?xt(O,v.classes):O},u?u.props.inputProps:{})},(y&&g||d)&&F==="outlined"?{notched:!0}:{},{ref:C,className:H(W.props.className,o,B.root)},!u&&{variant:F},A))})});rb.muiName="Select";const zA=rb,$A=ya.plugins,nb=b.createContext(),ob=()=>b.useContext(nb),UA=({children:e})=>{const[t,a]=b.useState("release"),[i,s]=b.useState($A);return m.jsx(nb.Provider,{value:{sortOption:t,setSortOption:a,sortedData:i,setSortedData:s},children:e})},Md=ya.plugins;function KA(){const{setSearchQuery:e,setIsSearchSubmitted:t}=nn(),{sortOption:a,setSortOption:i,setSortedData:s}=ob();b.useEffect(()=>{document.documentElement.style.scrollBehavior="auto",l(a),r()},[a]);function r(){var d=window.scrollY;window.onscroll=function(){var c=window.scrollY;d>c?document.querySelector("header").style.top="0":d>150&&(document.querySelector("header").style.top="-155px"),d=c}}function n(d){const c=Object.entries(d);return c.sort(([,p],[,u])=>u.commits_count-p.commits_count),Object.fromEntries(c)}function o(d){const c=Object.entries(d);return c.sort(([,p],[,u])=>!p.metadata.release_date&&!u.metadata.release_date?0:p.metadata.release_date?u.metadata.release_date?new Date(u.metadata.release_date)-new Date(p.metadata.release_date):-1:1),Object.fromEntries(c)}const l=d=>{i(d),e(""),t(!1);let c={};d==="commits"?c=n(Md):d=="alpha"?c=Md:d=="release"&&(c=o(Md)),s(c)};return m.jsx(e2,{children:m.jsxs(yS,{children:[m.jsx(h2,{children:"Sort"}),m.jsxs(zA,{value:a,label:"Sort",onChange:d=>l(d.target.value),children:[m.jsx(jd,{value:"commits",children:"Commits Count"}),m.jsx(jd,{value:"alpha",children:"Alphabetical"}),m.jsx(jd,{value:"release",children:"Recent Release"})]})]})})}var Xp={},VA=Zr;Object.defineProperty(Xp,"__esModule",{value:!0});var lb=Xp.default=void 0,HA=VA(Fp()),GA=m;lb=Xp.default=(0,HA.default)((0,GA.jsx)("path",{d:"M12 2C6.48 2 2 6.48 2 12s4.48 10 10 10 10-4.48 10-10S17.52 2 12 2m-2 15-5-5 1.41-1.41L10 14.17l7.59-7.59L19 8z"}),"CheckCircle");function XA(e){return pe("MuiDialog",e)}const JA=re("MuiDialog",["root","scrollPaper","scrollBody","container","paper","paperScrollPaper","paperScrollBody","paperWidthFalse","paperWidthXs","paperWidthSm","paperWidthMd","paperWidthLg","paperWidthXl","paperFullWidth","paperFullScreen"]),Ld=JA,QA=b.createContext({}),db=QA,YA=["aria-describedby","aria-labelledby","BackdropComponent","BackdropProps","children","className","disableEscapeKeyDown","fullScreen","fullWidth","maxWidth","onBackdropClick","onClick","onClose","open","PaperComponent","PaperProps","scroll","TransitionComponent","transitionDuration","TransitionProps"],ZA=z(Yv,{name:"MuiDialog",slot:"Backdrop",overrides:(e,t)=>t.backdrop})({zIndex:-1}),eR=e=>{const{classes:t,scroll:a,maxWidth:i,fullWidth:s,fullScreen:r}=e,n={root:["root"],container:["container",`scroll${X(a)}`],paper:["paper",`paperScroll${X(a)}`,`paperWidth${X(String(i))}`,s&&"paperFullWidth",r&&"paperFullScreen"]};return fe(n,XA,t)},tR=z(Kp,{name:"MuiDialog",slot:"Root",overridesResolver:(e,t)=>t.root})({"@media print":{position:"absolute !important"}}),aR=z("div",{name:"MuiDialog",slot:"Container",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.container,t[`scroll${X(a.scroll)}`]]}})(({ownerState:e})=>q({height:"100%","@media print":{height:"auto"},outline:0},e.scroll==="paper"&&{display:"flex",justifyContent:"center",alignItems:"center"},e.scroll==="body"&&{overflowY:"auto",overflowX:"hidden",textAlign:"center","&::after":{content:'""',display:"inline-block",verticalAlign:"middle",height:"100%",width:"0"}})),iR=z(un,{name:"MuiDialog",slot:"Paper",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.paper,t[`scrollPaper${X(a.scroll)}`],t[`paperWidth${X(String(a.maxWidth))}`],a.fullWidth&&t.paperFullWidth,a.fullScreen&&t.paperFullScreen]}})(({theme:e,ownerState:t})=>q({margin:32,position:"relative",overflowY:"auto","@media print":{overflowY:"visible",boxShadow:"none"}},t.scroll==="paper"&&{display:"flex",flexDirection:"column",maxHeight:"calc(100% - 64px)"},t.scroll==="body"&&{display:"inline-block",verticalAlign:"middle",textAlign:"left"},!t.maxWidth&&{maxWidth:"calc(100% - 64px)"},t.maxWidth==="xs"&&{maxWidth:e.breakpoints.unit==="px"?Math.max(e.breakpoints.values.xs,444):`max(${e.breakpoints.values.xs}${e.breakpoints.unit}, 444px)`,[`&.${Ld.paperScrollBody}`]:{[e.breakpoints.down(Math.max(e.breakpoints.values.xs,444)+32*2)]:{maxWidth:"calc(100% - 64px)"}}},t.maxWidth&&t.maxWidth!=="xs"&&{maxWidth:`${e.breakpoints.values[t.maxWidth]}${e.breakpoints.unit}`,[`&.${Ld.paperScrollBody}`]:{[e.breakpoints.down(e.breakpoints.values[t.maxWidth]+32*2)]:{maxWidth:"calc(100% - 64px)"}}},t.fullWidth&&{width:"calc(100% - 64px)"},t.fullScreen&&{margin:0,width:"100%",maxWidth:"100%",height:"100%",maxHeight:"none",borderRadius:0,[`&.${Ld.paperScrollBody}`]:{margin:0,maxWidth:"100%"}})),sR=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiDialog"}),s=on(),r={enter:s.transitions.duration.enteringScreen,exit:s.transitions.duration.leavingScreen},{"aria-describedby":n,"aria-labelledby":o,BackdropComponent:l,BackdropProps:d,children:c,className:p,disableEscapeKeyDown:u=!1,fullScreen:v=!1,fullWidth:f=!1,maxWidth:h="sm",onBackdropClick:T,onClick:y,onClose:g,open:_,PaperComponent:w=un,PaperProps:D={},scroll:x="paper",TransitionComponent:k=Qv,transitionDuration:S=r,TransitionProps:A}=i,P=U(i,YA),j=q({},i,{disableEscapeKeyDown:u,fullScreen:v,fullWidth:f,maxWidth:h,scroll:x}),E=eR(j),F=b.useRef(),I=C=>{F.current=C.target===C.currentTarget},B=C=>{y&&y(C),F.current&&(F.current=null,T&&T(C),g&&g(C,"backdropClick"))},O=yp(o),W=b.useMemo(()=>({titleId:O}),[O]);return m.jsx(tR,q({className:H(E.root,p),closeAfterTransition:!0,components:{Backdrop:ZA},componentsProps:{backdrop:q({transitionDuration:S,as:l},d)},disableEscapeKeyDown:u,onClose:g,open:_,ref:a,onClick:B,ownerState:j},P,{children:m.jsx(k,q({appear:!0,in:_,timeout:S,role:"presentation"},A,{children:m.jsx(aR,{className:H(E.container),onMouseDown:I,ownerState:j,children:m.jsx(iR,q({as:w,elevation:24,role:"dialog","aria-describedby":n,"aria-labelledby":O},D,{className:H(E.paper,D.className),ownerState:j,children:m.jsx(db.Provider,{value:W,children:c})}))})}))}))}),rR=sR;function nR(e){return pe("MuiDialogContent",e)}re("MuiDialogContent",["root","dividers"]);function oR(e){return pe("MuiDialogTitle",e)}const lR=re("MuiDialogTitle",["root"]),dR=lR,uR=["className","dividers"],cR=e=>{const{classes:t,dividers:a}=e;return fe({root:["root",a&&"dividers"]},nR,t)},pR=z("div",{name:"MuiDialogContent",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,a.dividers&&t.dividers]}})(({theme:e,ownerState:t})=>q({flex:"1 1 auto",WebkitOverflowScrolling:"touch",overflowY:"auto",padding:"20px 24px"},t.dividers?{padding:"16px 24px",borderTop:`1px solid ${(e.vars||e).palette.divider}`,borderBottom:`1px solid ${(e.vars||e).palette.divider}`}:{[`.${dR.root} + &`]:{paddingTop:0}})),fR=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiDialogContent"}),{className:s,dividers:r=!1}=i,n=U(i,uR),o=q({},i,{dividers:r}),l=cR(o);return m.jsx(pR,q({className:H(l.root,s),ownerState:o,ref:a},n))}),mR=fR;function hR(e){return pe("MuiDialogContentText",e)}re("MuiDialogContentText",["root"]);const gR=["children","className"],yR=e=>{const{classes:t}=e,i=fe({root:["root"]},hR,t);return q({},t,i)},_R=z(Gv,{shouldForwardProp:e=>It(e)||e==="classes",name:"MuiDialogContentText",slot:"Root",overridesResolver:(e,t)=>t.root})({}),vR=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiDialogContentText"}),{className:s}=i,r=U(i,gR),n=yR(r);return m.jsx(_R,q({component:"p",variant:"body1",color:"text.secondary",ref:a,ownerState:r,className:H(n.root,s)},i,{classes:n}))}),bR=vR,wR=["className","id"],TR=e=>{const{classes:t}=e;return fe({root:["root"]},oR,t)},kR=z(Gv,{name:"MuiDialogTitle",slot:"Root",overridesResolver:(e,t)=>t.root})({padding:"16px 24px",flex:"0 0 auto"}),qR=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiDialogTitle"}),{className:s,id:r}=i,n=U(i,wR),o=i,l=TR(o),{titleId:d=r}=b.useContext(db);return m.jsx(kR,q({component:"h2",className:H(l.root,s),ownerState:o,ref:a,variant:"h6",id:r??d},n))}),DR=qR,ub="data%3Aimage%2Fpng%3Bbase64%2CiVBORw0KGgoAAAANSUhEUgAAACMAAAAhCAYAAABTERJSAAAABHNCSVQICAgIfAhkiAAAAAlwSFlzAAAFhgAABYYBG6Yz4AAAABl0RVh0U29mdHdhcmUAd3d3Lmlua3NjYXBlLm9yZ5vuPBoAAAUbSURBVFiFzZhrbFRVEMd%2Fc%2B5uu6UUbIFC%2FUAUVEQCLbQJBIiBDyiImJiIhmohYNCkqJAQxASLF8tDgYRHBLXRhIcKNtFEhVDgAxBJqgmVh4JEKg3EIn2QYqBlt917xg%2BFss%2ByaDHOtzsz5z%2B%2FuZl7ztmF%2F5HJvxVQN6cPYX8%2FPLnOmsvNAvqfwuib%2FbNIk9cQeQnLcKRL5xLIV%2Fic9eJeunjPYbRs4FjQSpTB3aS1IpRKeeOOewajy%2FKKEO8Q0DuVdKy8IqsbPulxGHUfCBBu%2BwUYGuFuBTK7wQnht6PEbf4tlRomVRjCbXNjQEB0AyrFQOL5ENIJm7dTLZE6DPJCnEtFZVXDLny%2B4Sjv0PmmYu1ZdUek9RiMgoDmJ8V0L7XJqsZ3UW8YsBOwEeHeeFce7jEYXBy0m9m4BbXqSj2%2Bxnkg26MCVrN6DEZcwggtd8pTFx%2Fh3B9B50YLaFOPwXQKUt0tBLegtSomfBlfY13PwijbEnhztGzgJsK5h9W9qeWwBqjvyhB2iBs1Qz0AU974DciRGO8CVN8AJhAeMAdA3KbrKEtvxhsI%2B9emWiJlGBEU680Cfk%2BSsVqXZvcFYGXjF8ABVJ%2BTNfVXehyms1zzn1gmIOxLEB6E31%2FWBe5rnCarmo7elf7dJEeaLh80GasliI5F6Q9cAz1GY1OJVNDxTzQTw7iY%2FHEZRQY7xqJ9RU2LFe%2FYqakdP911ha0XhjjiTVAkDwgatWfCGeYocx8M3glG8g8EXhSrLrHnEFJ5Ymow%2FkhIYv6ttYUW1iFmEqqxdVoUs9FmsDYSqmtmJh3Cl1%2BVtl2s7owDUdocR5bceiyoSivGTT5vzpbzL1uoBpmcAAQgW7ArnKD9ng9rc%2BNgrobSNwpSkkhcRN%2BvmXLjIsDovYHHEfmsYFygPAnIDEQrQPzJYCOaLHLUfIt7Oq0LJn9fxkSgNCb1qEIQ5UKgT%2Fs6gJmVOOroJhQBXVqw118QtWLdyUxEP45sUpSzqP7RDdFYMyB9UReMiF1MzPwoUqHt8hjGFFeP5wZAbZ%2F0%2BcAtAAcji6LeSq%2FMYiAvSsdw3GtrfVSVFUBbIhwRWYR7yOcr%2FBi%2FB1MSJZ16JlgH1AGM3EO2QnmMyrSbTSiACgFBv4yCUapZkt9qwWVL7aeOyHvArJjm8%2Fz9BhdI4XcZgz2%2FvRALosjsk1ODOyMcJn9%2FYI6IrkS5vxMGdUwou2YKfyVqJpn5t9aNs3gbQMbdbkxnGdsr4bTHm2AxWo9yNZK4PXR3uzhAh%2BM0AZejnCrGdy0UvJxl0oMKgWSLR%2B1LH2aE9ViejiFs%2BXn6bTjng3MlIhJ1I1TkuLdg6OcAbD7Xx%2Bc3y9TrWAiSHqVkbZ2v9ilCo6s4AjwZCzFyD9mOL305nV9aonvsQeT2L0gVk4OwOJqXXVRW7naaxswDKVdlYLyMXAnntteYmws2xcVVZzq%2BtHPAooQggmJkc6TLSusOiL4RKgwzzYU1iFQgiUBA1H7E8yPau%2BZl9P7AblVNebtHqTgxLfRqrNvZWjsHZFuqMqKcDWdlFjF7UGvX8Jn24DyEAykJwNcdg0OvJ4p5pQ9tV6SMlP4A0PNh8aYze1ArROyUNTNouy8tNF3Rt0CSXb6bRFl4%2FIfQzNMjaE9WwpYOWQnOdEF%2BTdJNO0iFh7%2BI0kfORzQZb6P2kymS9oTxzBiM9rUqLWr1WE5G6ODhycQd%2FUnNVeMbcH68hYkGycNoUNWc8fxaxfwhDbHpfwM5oeTY7rUX8QAAAABJRU5ErkJggg%3D%3D";const Ih=ya.status_dict,xR={}.VITE_PR_PREVIEW_PATH||"/aiida-registry/";function CR(){const{sortOption:e,sortedData:t}=ob();return m.jsx(m.Fragment,{children:Object.entries(t).map(([a,i])=>m.jsxs("div",{className:"submenu-entry",children:[m.jsx(Bi,{to:`/${a}`,children:m.jsxs("h2",{style:{display:"inline"},children:[a," "]})}),i.is_installable==="True"&&m.jsx(SR,{}),m.jsxs("p",{className:"currentstate",children:[m.jsx("img",{className:"svg-badge",src:`${xR}${Ih[i.development_status][1]}`,title:Ih[i.development_status][0]})," ",i.aiida_version&&m.jsx("img",{className:"svg-badge",title:`Compatible with aiida-core ${i.aiida_version}`,src:`https://img.shields.io/badge/AiiDA-${i.aiida_version}-007ec6.svg?logo=${ub}`}),e==="commits"&&m.jsx("img",{className:"svg-badge",style:{padding:"3px"},src:`https://img.shields.io/badge/Yearly%20Commits-${i.commits_count}-007ec6.svg`}),e==="release"&&i.metadata.release_date&&m.jsx("img",{className:"svg-badge",style:{padding:"3px"},src:`https://img.shields.io/badge/Recent%20Release-${i.metadata.release_date.replace(/-/g,"/")}-007ec6.svg`})]}),m.jsx("p",{children:i.metadata.description}),m.jsxs("ul",{className:"plugin-info",children:[m.jsx("li",{children:m.jsx("a",{href:i.code_home,children:"Source Code"})}),i.documentation_url&&m.jsx("li",{children:m.jsx("a",{href:i.documentation_url,children:"Documentation"})}),m.jsx("li",{children:m.jsx(Bi,{to:`/${a}`,children:"Plugin details"})})]}),i.summaryinfo&&m.jsx(m.Fragment,{children:m.jsx("p",{className:"summaryinfo",children:i.summaryinfo.map(s=>m.jsxs("span",{className:"badge",children:[m.jsx("span",{className:`badge-left ${s.colorclass}`,children:s.text}),m.jsx("span",{className:"badge-right",children:s.count})]},s.text))})})]},a))})}function SR(){const[e,t]=b.useState(!1),a=()=>{t(!0)},i=()=>{t(!1)};return m.jsxs(m.Fragment,{children:[m.jsxs("div",{className:"classbox",style:{backgroundColor:"transparent"},children:[m.jsx(lb,{onClick:a,style:{color:"green",cursor:"pointer",marginBottom:"-5"}}),m.jsx("span",{className:"tooltiptext",children:"Plugin successfully installed"})]}),m.jsxs(rR,{open:e,onClose:i,children:[m.jsx(DR,{children:"This plugin can be installed with the latest aiida-core version."}),m.jsx(mR,{children:m.jsxs(bR,{children:["This check mark indicates that this plugin was installed successfully inside the latest",m.jsxs("a",{rel:"noopener noreferrer",target:"_blank",href:"https://hub.docker.com/r/aiidateam/aiida-core-with-services",children:[m.jsx("code",{children:"aiida-core-with-services"})," docker image"]}),". For in-depth compatibility tests see ",m.jsx("a",{rel:"noopener noreferrer",target:"_blank",href:"https://github.com/aiidateam/aiida-registry/tree/master#how-to-fix-registry-warnings-and-errors",children:"how to reproduce it locally"}),"."]})})]})]})}const PR=ya.globalsummary,AR=ya.plugins,RR=Object.keys(AR).length;function IR(){return m.jsxs(m.Fragment,{children:[m.jsxs("h2",{children:["Registered plugin packages: ",RR]}),m.jsx("div",{className:"globalsummary-box",children:m.jsx("div",{style:{display:"table"},children:PR.map(e=>m.jsxs("span",{className:"badge",style:{display:"table-row",lineHeight:2},children:[m.jsx("span",{style:{display:"table-cell",float:"none",textAlign:"right"},children:m.jsxs("span",{className:`badge-left ${e.colorclass} tooltip`,style:{float:"none",display:"inline",textAlign:"right",border:"none"},children:[e.name,e.tooltip&&m.jsx("span",{className:"tooltiptext",children:e.tooltip})]})}),m.jsx("span",{style:{display:"table-cell",float:"none",textAlign:"left"},children:m.jsxs("span",{className:"badge-right",style:{float:"none",display:"inline",textAlign:"left",border:"none"},children:[e.total_num," plugin",e.total_num!==1?"s":""," in ",e.num_entries," package",e.num_entries!==1?"s":""]})})]},e.name))})})]})}function NR(){const{isSearchSubmitted:e}=nn();return m.jsxs("main",{className:"fade-enter",children:[m.jsx(IR,{}),m.jsxs("div",{id:"entrylist",children:[m.jsx("h1",{children:"Package list"}),m.jsxs("div",{className:"bar-container",children:[m.jsx("div",{style:{flex:"1",marginRight:"10px"},children:m.jsx(LC,{})}),m.jsx(KA,{})]}),e===!0?m.jsx(WC,{}):m.jsx(CR,{})]})]})}function ki(){return ki=Object.assign?Object.assign.bind():function(e){for(var t=1;t(e[t.toLowerCase()]=t,e),{class:"className",for:"htmlFor"}),Eh={amp:"&",apos:"'",gt:">",lt:"<",nbsp:" ",quot:"“"},ER=["style","script"],FR=/([-A-Z0-9_:]+)(?:\s*=\s*(?:(?:"((?:\\.|[^"])*)")|(?:'((?:\\.|[^'])*)')|(?:\{((?:\\.|{[^}]*?}|[^}])*)\})))?/gi,BR=/mailto:/i,OR=/\n{2,}$/,cb=/^(\s*>[\s\S]*?)(?=\n\n|$)/,MR=/^ *> ?/gm,LR=/^(?:\[!([^\]]*)\]\n)?([\s\S]*)/,WR=/^ {2,}\n/,zR=/^(?:( *[-*_])){3,} *(?:\n *)+\n/,pb=/^(?: {1,3})?(`{3,}|~{3,}) *(\S+)? *([^\n]*?)?\n([\s\S]*?)(?:\1\n?|$)/,fb=/^(?: {4}[^\n]+\n*)+(?:\n *)+\n?/,$R=/^(`+)\s*([\s\S]*?[^`])\s*\1(?!`)/,UR=/^(?:\n *)*\n/,KR=/\r\n?/g,VR=/^\[\^([^\]]+)](:(.*)((\n+ {4,}.*)|(\n(?!\[\^).+))*)/,HR=/^\[\^([^\]]+)]/,GR=/\f/g,XR=/^---[ \t]*\n(.|\n)*\n---[ \t]*\n/,JR=/^\s*?\[(x|\s)\]/,mb=/^ *(#{1,6}) *([^\n]+?)(?: +#*)?(?:\n *)*(?:\n|$)/,hb=/^ *(#{1,6}) +([^\n]+?)(?: +#*)?(?:\n *)*(?:\n|$)/,gb=/^([^\n]+)\n *(=|-){3,} *(?:\n *)+\n/,cc=/^ *(?!<[a-z][^ >/]* ?\/>)<([a-z][^ >/]*) ?((?:[^>]*[^/])?)>\n?(\s*(?:<\1[^>]*?>[\s\S]*?<\/\1>|(?!<\1\b)[\s\S])*?)<\/\1>(?!<\/\1>)\n*/i,QR=/&([a-z0-9]+|#[0-9]{1,6}|#x[0-9a-fA-F]{1,6});/gi,yb=/^)/,YR=/^(data|aria|x)-[a-z_][a-z\d_.-]*$/,pc=/^ *<([a-z][a-z0-9:]*)(?:\s+((?:<.*?>|[^>])*))?\/?>(?!<\/\1>)(\s*\n)?/i,ZR=/^\{.*\}$/,e3=/^(https?:\/\/[^\s<]+[^<.,:;"')\]\s])/,t3=/^<([^ >]+@[^ >]+)>/,a3=/^<([^ >]+:\/[^ >]+)>/,i3=/-([a-z])?/gi,_b=/^(\|.*)\n(?: *(\|? *[-:]+ *\|[-| :]*)\n((?:.*\|.*\n)*))?\n?/,s3=/^\[([^\]]*)\]:\s+]+)>?\s*("([^"]*)")?/,r3=/^!\[([^\]]*)\] ?\[([^\]]*)\]/,n3=/^\[([^\]]*)\] ?\[([^\]]*)\]/,o3=/(\n|^[-*]\s|^#|^ {2,}|^-{2,}|^>\s)/,l3=/\t/g,d3=/(^ *\||\| *$)/g,u3=/^ *:-+: *$/,c3=/^ *:-+ *$/,p3=/^ *-+: *$/,Vl="((?:\\[.*?\\][([].*?[)\\]]|<.*?>(?:.*?<.*?>)?|`.*?`|~~.*?~~|==.*?==|.|\\n)*?)",f3=new RegExp(`^([*_])\\1${Vl}\\1\\1(?!\\1)`),m3=new RegExp(`^([*_])${Vl}\\1(?!\\1|\\w)`),h3=new RegExp(`^==${Vl}==`),g3=new RegExp(`^~~${Vl}~~`),y3=/^\\([^0-9A-Za-z\s])/,_3=/^[\s\S]+?(?=[^0-9A-Z\s\u00c0-\uffff&#;.()'"]|\d+\.|\n\n| {2,}\n|\w+:\S|$)/i,v3=/^\n+/,b3=/^([ \t]*)/,w3=/\\([^\\])/g,Fh=/ *\n+$/,T3=/(?:^|\n)( *)$/,Jp="(?:\\d+\\.)",Qp="(?:[*+-])";function vb(e){return"( *)("+(e===1?Jp:Qp)+") +"}const bb=vb(1),wb=vb(2);function Tb(e){return new RegExp("^"+(e===1?bb:wb))}const k3=Tb(1),q3=Tb(2);function kb(e){return new RegExp("^"+(e===1?bb:wb)+"[^\\n]*(?:\\n(?!\\1"+(e===1?Jp:Qp)+" )[^\\n]*)*(\\n|$)","gm")}const qb=kb(1),Db=kb(2);function xb(e){const t=e===1?Jp:Qp;return new RegExp("^( *)("+t+") [\\s\\S]+?(?:\\n{2,}(?! )(?!\\1"+t+" (?!"+t+" ))\\n*|\\s*\\n*$)")}const Cb=xb(1),Sb=xb(2);function Bh(e,t){const a=t===1,i=a?Cb:Sb,s=a?qb:Db,r=a?k3:q3;return{match(n,o){const l=T3.exec(o.prevCapture);return l&&(o.list||!o.inline&&!o.simple)?i.exec(n=l[1]+n):null},order:1,parse(n,o,l){const d=a?+n[2]:void 0,c=n[0].replace(OR,` +`).match(s);let p=!1;return{items:c.map(function(u,v){const f=r.exec(u)[0].length,h=new RegExp("^ {1,"+f+"}","gm"),T=u.replace(h,"").replace(r,""),y=v===c.length-1,g=T.indexOf(` + +`)!==-1||y&&p;p=g;const _=l.inline,w=l.list;let D;l.list=!0,g?(l.inline=!1,D=T.replace(Fh,` + +`)):(l.inline=!0,D=T.replace(Fh,""));const x=o(D,l);return l.inline=_,l.list=w,x}),ordered:a,start:d}},render:(n,o,l)=>e(n.ordered?"ol":"ul",{key:l.key,start:n.type===$.orderedList?n.start:void 0},n.items.map(function(d,c){return e("li",{key:c},o(d,l))}))}}const D3=new RegExp(`^\\[((?:\\[[^\\]]*\\]|[^\\[\\]]|\\](?=[^\\[]*\\]))*)\\]\\(\\s*?(?:\\s+['"]([\\s\\S]*?)['"])?\\s*\\)`),x3=/^!\[(.*?)\]\( *((?:\([^)]*\)|[^() ])*) *"?([^)"]*)?"?\)/,Pb=[cb,pb,fb,mb,gb,hb,yb,_b,qb,Cb,Db,Sb],C3=[...Pb,/^[^\n]+(?: \n|\n{2,})/,cc,pc];function ar(e){return e.replace(/[ÀÁÂÃÄÅàáâãäåæÆ]/g,"a").replace(/[çÇ]/g,"c").replace(/[ðÐ]/g,"d").replace(/[ÈÉÊËéèêë]/g,"e").replace(/[ÏïÎîÍíÌì]/g,"i").replace(/[Ññ]/g,"n").replace(/[øØœŒÕõÔôÓóÒò]/g,"o").replace(/[ÜüÛûÚúÙù]/g,"u").replace(/[ŸÿÝý]/g,"y").replace(/[^a-z0-9- ]/gi,"").replace(/ /gi,"-").toLowerCase()}function S3(e){return p3.test(e)?"right":u3.test(e)?"center":c3.test(e)?"left":null}function Oh(e,t,a,i){const s=a.inTable;a.inTable=!0;let r=e.trim().split(/( *(?:`[^`]*`|\\\||\|) *)/).reduce((o,l)=>(l.trim()==="|"?o.push(i?{type:$.tableSeparator}:{type:$.text,text:l}):l!==""&&o.push.apply(o,t(l,a)),o),[]);a.inTable=s;let n=[[]];return r.forEach(function(o,l){o.type===$.tableSeparator?l!==0&&l!==r.length-1&&n.push([]):(o.type!==$.text||r[l+1]!=null&&r[l+1].type!==$.tableSeparator||(o.text=o.text.trimEnd()),n[n.length-1].push(o))}),n}function P3(e,t,a){a.inline=!0;const i=e[2]?e[2].replace(d3,"").split("|").map(S3):[],s=e[3]?function(n,o,l){return n.trim().split(` +`).map(function(d){return Oh(d,o,l,!0)})}(e[3],t,a):[],r=Oh(e[1],t,a,!!s.length);return a.inline=!1,s.length?{align:i,cells:s,header:r,type:$.table}:{children:r,type:$.paragraph}}function Mh(e,t){return e.align[t]==null?{}:{textAlign:e.align[t]}}function $a(e){return function(t,a){return a.inline?e.exec(t):null}}function Ua(e){return function(t,a){return a.inline||a.simple?e.exec(t):null}}function wa(e){return function(t,a){return a.inline||a.simple?null:e.exec(t)}}function ir(e){return function(t){return e.exec(t)}}function A3(e,t){if(t.inline||t.simple)return null;let a="";e.split(` +`).every(s=>!Pb.some(r=>r.test(s))&&(a+=s+` +`,s.trim()));const i=a.trimEnd();return i==""?null:[a,i]}function R3(e){try{if(decodeURIComponent(e).replace(/[^A-Za-z0-9/:]/g,"").match(/^\s*(javascript|vbscript|data(?!:image)):/i))return null}catch{return null}return e}function Lh(e){return e.replace(w3,"$1")}function io(e,t,a){const i=a.inline||!1,s=a.simple||!1;a.inline=!0,a.simple=!0;const r=e(t,a);return a.inline=i,a.simple=s,r}function I3(e,t,a){const i=a.inline||!1,s=a.simple||!1;a.inline=!1,a.simple=!0;const r=e(t,a);return a.inline=i,a.simple=s,r}function N3(e,t,a){const i=a.inline||!1;a.inline=!1;const s=e(t,a);return a.inline=i,s}const Wd=(e,t,a)=>({children:io(t,e[1],a)});function zd(){return{}}function $d(){return null}function j3(...e){return e.filter(Boolean).join(" ")}function Ud(e,t,a){let i=e;const s=t.split(".");for(;s.length&&(i=i[s[0]],i!==void 0);)s.shift();return i||a}function E3(e="",t={}){function a(u,v,...f){const h=Ud(t.overrides,`${u}.props`,{});return t.createElement(function(T,y){const g=Ud(y,T);return g?typeof g=="function"||typeof g=="object"&&"render"in g?g:Ud(y,`${T}.component`,T):T}(u,t.overrides),ki({},v,h,{className:j3(v==null?void 0:v.className,h.className)||void 0}),...f)}function i(u){u=u.replace(XR,"");let v=!1;t.forceInline?v=!0:t.forceBlock||(v=o3.test(u)===!1);const f=d(l(v?u:`${u.trimEnd().replace(v3,"")} + +`,{inline:v}));for(;typeof f[f.length-1]=="string"&&!f[f.length-1].trim();)f.pop();if(t.wrapper===null)return f;const h=t.wrapper||(v?"span":"div");let T;if(f.length>1||t.forceWrapper)T=f;else{if(f.length===1)return T=f[0],typeof T=="string"?a("span",{key:"outer"},T):T;T=null}return b.createElement(h,{key:"outer"},T)}function s(u,v){const f=v.match(FR);return f?f.reduce(function(h,T,y){const g=T.indexOf("=");if(g!==-1){const _=function(k){return k.indexOf("-")!==-1&&k.match(YR)===null&&(k=k.replace(i3,function(S,A){return A.toUpperCase()})),k}(T.slice(0,g)).trim(),w=function(k){const S=k[0];return(S==='"'||S==="'")&&k.length>=2&&k[k.length-1]===S?k.slice(1,-1):k}(T.slice(g+1).trim()),D=jh[_]||_;if(D==="ref")return h;const x=h[D]=function(k,S,A,P){return S==="style"?A.split(/;\s?/).reduce(function(j,E){const F=E.slice(0,E.indexOf(":"));return j[F.trim().replace(/(-[a-z])/g,I=>I[1].toUpperCase())]=E.slice(F.length+1).trim(),j},{}):S==="href"||S==="src"?P(A,k,S):(A.match(ZR)&&(A=A.slice(1,A.length-1)),A==="true"||A!=="false"&&A)}(u,_,w,t.sanitizer);typeof x=="string"&&(cc.test(x)||pc.test(x))&&(h[D]=b.cloneElement(i(x.trim()),{key:y}))}else T!=="style"&&(h[jh[T]||T]=!0);return h},{}):null}t.overrides=t.overrides||{},t.sanitizer=t.sanitizer||R3,t.slugify=t.slugify||ar,t.namedCodesToUnicode=t.namedCodesToUnicode?ki({},Eh,t.namedCodesToUnicode):Eh,t.createElement=t.createElement||b.createElement;const r=[],n={},o={[$.blockQuote]:{match:wa(cb),order:1,parse(u,v,f){const[,h,T]=u[0].replace(MR,"").match(LR);return{alert:h,children:v(T,f)}},render(u,v,f){const h={key:f.key};return u.alert&&(h.className="markdown-alert-"+t.slugify(u.alert.toLowerCase(),ar),u.children.unshift({attrs:{},children:[{type:$.text,text:u.alert}],noInnerParse:!0,type:$.htmlBlock,tag:"header"})),a("blockquote",h,v(u.children,f))}},[$.breakLine]:{match:ir(WR),order:1,parse:zd,render:(u,v,f)=>a("br",{key:f.key})},[$.breakThematic]:{match:wa(zR),order:1,parse:zd,render:(u,v,f)=>a("hr",{key:f.key})},[$.codeBlock]:{match:wa(fb),order:0,parse:u=>({lang:void 0,text:u[0].replace(/^ {4}/gm,"").replace(/\n+$/,"")}),render:(u,v,f)=>a("pre",{key:f.key},a("code",ki({},u.attrs,{className:u.lang?`lang-${u.lang}`:""}),u.text))},[$.codeFenced]:{match:wa(pb),order:0,parse:u=>({attrs:s("code",u[3]||""),lang:u[2]||void 0,text:u[4],type:$.codeBlock})},[$.codeInline]:{match:Ua($R),order:3,parse:u=>({text:u[2]}),render:(u,v,f)=>a("code",{key:f.key},u.text)},[$.footnote]:{match:wa(VR),order:0,parse:u=>(r.push({footnote:u[2],identifier:u[1]}),{}),render:$d},[$.footnoteReference]:{match:$a(HR),order:1,parse:u=>({target:`#${t.slugify(u[1],ar)}`,text:u[1]}),render:(u,v,f)=>a("a",{key:f.key,href:t.sanitizer(u.target,"a","href")},a("sup",{key:f.key},u.text))},[$.gfmTask]:{match:$a(JR),order:1,parse:u=>({completed:u[1].toLowerCase()==="x"}),render:(u,v,f)=>a("input",{checked:u.completed,key:f.key,readOnly:!0,type:"checkbox"})},[$.heading]:{match:wa(t.enforceAtxHeadings?hb:mb),order:1,parse:(u,v,f)=>({children:io(v,u[2],f),id:t.slugify(u[2],ar),level:u[1].length}),render:(u,v,f)=>a(`h${u.level}`,{id:u.id,key:f.key},v(u.children,f))},[$.headingSetext]:{match:wa(gb),order:0,parse:(u,v,f)=>({children:io(v,u[1],f),level:u[2]==="="?1:2,type:$.heading})},[$.htmlBlock]:{match:ir(cc),order:1,parse(u,v,f){const[,h]=u[3].match(b3),T=new RegExp(`^${h}`,"gm"),y=u[3].replace(T,""),g=(_=y,C3.some(S=>S.test(_))?N3:io);var _;const w=u[1].toLowerCase(),D=ER.indexOf(w)!==-1,x=(D?w:u[1]).trim(),k={attrs:s(x,u[2]),noInnerParse:D,tag:x};return f.inAnchor=f.inAnchor||w==="a",D?k.text=u[3]:k.children=g(v,y,f),f.inAnchor=!1,k},render:(u,v,f)=>a(u.tag,ki({key:f.key},u.attrs),u.text||(u.children?v(u.children,f):""))},[$.htmlSelfClosing]:{match:ir(pc),order:1,parse(u){const v=u[1].trim();return{attrs:s(v,u[2]||""),tag:v}},render:(u,v,f)=>a(u.tag,ki({},u.attrs,{key:f.key}))},[$.htmlComment]:{match:ir(yb),order:1,parse:()=>({}),render:$d},[$.image]:{match:Ua(x3),order:1,parse:u=>({alt:u[1],target:Lh(u[2]),title:u[3]}),render:(u,v,f)=>a("img",{key:f.key,alt:u.alt||void 0,title:u.title||void 0,src:t.sanitizer(u.target,"img","src")})},[$.link]:{match:$a(D3),order:3,parse:(u,v,f)=>({children:I3(v,u[1],f),target:Lh(u[2]),title:u[3]}),render:(u,v,f)=>a("a",{key:f.key,href:t.sanitizer(u.target,"a","href"),title:u.title},v(u.children,f))},[$.linkAngleBraceStyleDetector]:{match:$a(a3),order:0,parse:u=>({children:[{text:u[1],type:$.text}],target:u[1],type:$.link})},[$.linkBareUrlDetector]:{match:(u,v)=>v.inAnchor||t.disableAutoLink?null:$a(e3)(u,v),order:0,parse:u=>({children:[{text:u[1],type:$.text}],target:u[1],title:void 0,type:$.link})},[$.linkMailtoDetector]:{match:$a(t3),order:0,parse(u){let v=u[1],f=u[1];return BR.test(f)||(f="mailto:"+f),{children:[{text:v.replace("mailto:",""),type:$.text}],target:f,type:$.link}}},[$.orderedList]:Bh(a,1),[$.unorderedList]:Bh(a,2),[$.newlineCoalescer]:{match:wa(UR),order:3,parse:zd,render:()=>` +`},[$.paragraph]:{match:A3,order:3,parse:Wd,render:(u,v,f)=>a("p",{key:f.key},v(u.children,f))},[$.ref]:{match:$a(s3),order:0,parse:u=>(n[u[1]]={target:u[2],title:u[4]},{}),render:$d},[$.refImage]:{match:Ua(r3),order:0,parse:u=>({alt:u[1]||void 0,ref:u[2]}),render:(u,v,f)=>n[u.ref]?a("img",{key:f.key,alt:u.alt,src:t.sanitizer(n[u.ref].target,"img","src"),title:n[u.ref].title}):null},[$.refLink]:{match:$a(n3),order:0,parse:(u,v,f)=>({children:v(u[1],f),fallbackChildren:u[0],ref:u[2]}),render:(u,v,f)=>n[u.ref]?a("a",{key:f.key,href:t.sanitizer(n[u.ref].target,"a","href"),title:n[u.ref].title},v(u.children,f)):a("span",{key:f.key},u.fallbackChildren)},[$.table]:{match:wa(_b),order:1,parse:P3,render(u,v,f){const h=u;return a("table",{key:f.key},a("thead",null,a("tr",null,h.header.map(function(T,y){return a("th",{key:y,style:Mh(h,y)},v(T,f))}))),a("tbody",null,h.cells.map(function(T,y){return a("tr",{key:y},T.map(function(g,_){return a("td",{key:_,style:Mh(h,_)},v(g,f))}))})))}},[$.text]:{match:ir(_3),order:4,parse:u=>({text:u[0].replace(QR,(v,f)=>t.namedCodesToUnicode[f]?t.namedCodesToUnicode[f]:v)}),render:u=>u.text},[$.textBolded]:{match:Ua(f3),order:2,parse:(u,v,f)=>({children:v(u[2],f)}),render:(u,v,f)=>a("strong",{key:f.key},v(u.children,f))},[$.textEmphasized]:{match:Ua(m3),order:3,parse:(u,v,f)=>({children:v(u[2],f)}),render:(u,v,f)=>a("em",{key:f.key},v(u.children,f))},[$.textEscaped]:{match:Ua(y3),order:1,parse:u=>({text:u[1],type:$.text})},[$.textMarked]:{match:Ua(h3),order:3,parse:Wd,render:(u,v,f)=>a("mark",{key:f.key},v(u.children,f))},[$.textStrikethroughed]:{match:Ua(g3),order:3,parse:Wd,render:(u,v,f)=>a("del",{key:f.key},v(u.children,f))}};t.disableParsingRawHTML===!0&&(delete o[$.htmlBlock],delete o[$.htmlSelfClosing]);const l=function(u){let v=Object.keys(u);function f(h,T){let y=[];for(T.prevCapture=T.prevCapture||"";h;){let g=0;for(;gy(f,h,T),f,h,T):y(f,h,T)}}(o,t.renderRule),function u(v,f={}){if(Array.isArray(v)){const h=f.key,T=[];let y=!1;for(let g=0;g{let{children:t="",options:a}=e,i=function(s,r){if(s==null)return{};var n,o,l={},d=Object.keys(s);for(o=0;o=0||(l[n]=s[n]);return l}(e,jR);return b.cloneElement(E3(t,a),i)};const F3=["className","elementType","ownerState","externalForwardedProps","getSlotOwnerState","internalForwardedProps"],B3=["component","slots","slotProps"],O3=["component"];function Wh(e,t){const{className:a,elementType:i,ownerState:s,externalForwardedProps:r,getSlotOwnerState:n,internalForwardedProps:o}=t,l=U(t,F3),{component:d,slots:c={[e]:void 0},slotProps:p={[e]:void 0}}=r,u=U(r,B3),v=c[e]||i,f=U_(p[e],s),h=$_(q({className:a},l,{externalForwardedProps:e==="root"?u:void 0,externalSlotProps:f})),{props:{component:T},internalRef:y}=h,g=U(h.props,O3),_=Ke(y,f==null?void 0:f.ref,t.ref),w=n?n(g):{},D=q({},s,w),x=e==="root"?T||d:T,k=L_(v,q({},e==="root"&&!d&&!c[e]&&o,e!=="root"&&!c[e]&&o,g,x&&{as:x},{ref:_}),D);return Object.keys(w).forEach(S=>{delete k[S]}),[v,k]}function M3(e){return pe("MuiAlert",e)}const L3=re("MuiAlert",["root","action","icon","message","filled","colorSuccess","colorInfo","colorWarning","colorError","filledSuccess","filledInfo","filledWarning","filledError","outlined","outlinedSuccess","outlinedInfo","outlinedWarning","outlinedError","standard","standardSuccess","standardInfo","standardWarning","standardError"]),zh=L3;function W3(e){return pe("MuiIconButton",e)}const z3=re("MuiIconButton",["root","disabled","colorInherit","colorPrimary","colorSecondary","colorError","colorInfo","colorSuccess","colorWarning","edgeStart","edgeEnd","sizeSmall","sizeMedium","sizeLarge"]),$3=z3,U3=["edge","children","className","color","disabled","disableFocusRipple","size"],K3=e=>{const{classes:t,disabled:a,color:i,edge:s,size:r}=e,n={root:["root",a&&"disabled",i!=="default"&&`color${X(i)}`,s&&`edge${X(s)}`,`size${X(r)}`]};return fe(n,W3,t)},V3=z(Vv,{name:"MuiIconButton",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,a.color!=="default"&&t[`color${X(a.color)}`],a.edge&&t[`edge${X(a.edge)}`],t[`size${X(a.size)}`]]}})(({theme:e,ownerState:t})=>q({textAlign:"center",flex:"0 0 auto",fontSize:e.typography.pxToRem(24),padding:8,borderRadius:"50%",overflow:"visible",color:(e.vars||e).palette.action.active,transition:e.transitions.create("background-color",{duration:e.transitions.duration.shortest})},!t.disableRipple&&{"&:hover":{backgroundColor:e.vars?`rgba(${e.vars.palette.action.activeChannel} / ${e.vars.palette.action.hoverOpacity})`:xa(e.palette.action.active,e.palette.action.hoverOpacity),"@media (hover: none)":{backgroundColor:"transparent"}}},t.edge==="start"&&{marginLeft:t.size==="small"?-3:-12},t.edge==="end"&&{marginRight:t.size==="small"?-3:-12}),({theme:e,ownerState:t})=>{var a;const i=(a=(e.vars||e).palette)==null?void 0:a[t.color];return q({},t.color==="inherit"&&{color:"inherit"},t.color!=="inherit"&&t.color!=="default"&&q({color:i==null?void 0:i.main},!t.disableRipple&&{"&:hover":q({},i&&{backgroundColor:e.vars?`rgba(${i.mainChannel} / ${e.vars.palette.action.hoverOpacity})`:xa(i.main,e.palette.action.hoverOpacity)},{"@media (hover: none)":{backgroundColor:"transparent"}})}),t.size==="small"&&{padding:5,fontSize:e.typography.pxToRem(18)},t.size==="large"&&{padding:12,fontSize:e.typography.pxToRem(28)},{[`&.${$3.disabled}`]:{backgroundColor:"transparent",color:(e.vars||e).palette.action.disabled}})}),H3=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiIconButton"}),{edge:s=!1,children:r,className:n,color:o="default",disabled:l=!1,disableFocusRipple:d=!1,size:c="medium"}=i,p=U(i,U3),u=q({},i,{edge:s,color:o,disabled:l,disableFocusRipple:d,size:c}),v=K3(u);return m.jsx(V3,q({className:H(v.root,n),centerRipple:!0,focusRipple:!d,disabled:l,ref:a},p,{ownerState:u,children:r}))}),G3=H3,X3=zi(m.jsx("path",{d:"M20,12A8,8 0 0,1 12,20A8,8 0 0,1 4,12A8,8 0 0,1 12,4C12.76,4 13.5,4.11 14.2, 4.31L15.77,2.74C14.61,2.26 13.34,2 12,2A10,10 0 0,0 2,12A10,10 0 0,0 12,22A10,10 0 0, 0 22,12M7.91,10.08L6.5,11.5L11,16L21,6L19.59,4.58L11,13.17L7.91,10.08Z"}),"SuccessOutlined"),J3=zi(m.jsx("path",{d:"M12 5.99L19.53 19H4.47L12 5.99M12 2L1 21h22L12 2zm1 14h-2v2h2v-2zm0-6h-2v4h2v-4z"}),"ReportProblemOutlined"),Q3=zi(m.jsx("path",{d:"M11 15h2v2h-2zm0-8h2v6h-2zm.99-5C6.47 2 2 6.48 2 12s4.47 10 9.99 10C17.52 22 22 17.52 22 12S17.52 2 11.99 2zM12 20c-4.42 0-8-3.58-8-8s3.58-8 8-8 8 3.58 8 8-3.58 8-8 8z"}),"ErrorOutline"),Y3=zi(m.jsx("path",{d:"M11,9H13V7H11M12,20C7.59,20 4,16.41 4,12C4,7.59 7.59,4 12,4C16.41,4 20,7.59 20, 12C20,16.41 16.41,20 12,20M12,2A10,10 0 0,0 2,12A10,10 0 0,0 12,22A10,10 0 0,0 22,12A10, 10 0 0,0 12,2M11,17H13V11H11V17Z"}),"InfoOutlined"),Z3=zi(m.jsx("path",{d:"M19 6.41L17.59 5 12 10.59 6.41 5 5 6.41 10.59 12 5 17.59 6.41 19 12 13.41 17.59 19 19 17.59 13.41 12z"}),"Close"),eI=["action","children","className","closeText","color","components","componentsProps","icon","iconMapping","onClose","role","severity","slotProps","slots","variant"],tI=e=>{const{variant:t,color:a,severity:i,classes:s}=e,r={root:["root",`color${X(a||i)}`,`${t}${X(a||i)}`,`${t}`],icon:["icon"],message:["message"],action:["action"]};return fe(r,M3,s)},aI=z(un,{name:"MuiAlert",slot:"Root",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.root,t[a.variant],t[`${a.variant}${X(a.color||a.severity)}`]]}})(({theme:e})=>{const t=e.palette.mode==="light"?Yu:Zu,a=e.palette.mode==="light"?Zu:Yu;return q({},e.typography.body2,{backgroundColor:"transparent",display:"flex",padding:"6px 16px",variants:[...Object.entries(e.palette).filter(([,i])=>i.main&&i.light).map(([i])=>({props:{colorSeverity:i,variant:"standard"},style:{color:e.vars?e.vars.palette.Alert[`${i}Color`]:t(e.palette[i].light,.6),backgroundColor:e.vars?e.vars.palette.Alert[`${i}StandardBg`]:a(e.palette[i].light,.9),[`& .${zh.icon}`]:e.vars?{color:e.vars.palette.Alert[`${i}IconColor`]}:{color:e.palette[i].main}}})),...Object.entries(e.palette).filter(([,i])=>i.main&&i.light).map(([i])=>({props:{colorSeverity:i,variant:"outlined"},style:{color:e.vars?e.vars.palette.Alert[`${i}Color`]:t(e.palette[i].light,.6),border:`1px solid ${(e.vars||e).palette[i].light}`,[`& .${zh.icon}`]:e.vars?{color:e.vars.palette.Alert[`${i}IconColor`]}:{color:e.palette[i].main}}})),...Object.entries(e.palette).filter(([,i])=>i.main&&i.dark).map(([i])=>({props:{colorSeverity:i,variant:"filled"},style:q({fontWeight:e.typography.fontWeightMedium},e.vars?{color:e.vars.palette.Alert[`${i}FilledColor`],backgroundColor:e.vars.palette.Alert[`${i}FilledBg`]}:{backgroundColor:e.palette.mode==="dark"?e.palette[i].dark:e.palette[i].main,color:e.palette.getContrastText(e.palette[i].main)})}))]})}),iI=z("div",{name:"MuiAlert",slot:"Icon",overridesResolver:(e,t)=>t.icon})({marginRight:12,padding:"7px 0",display:"flex",fontSize:22,opacity:.9}),sI=z("div",{name:"MuiAlert",slot:"Message",overridesResolver:(e,t)=>t.message})({padding:"8px 0",minWidth:0,overflow:"auto"}),$h=z("div",{name:"MuiAlert",slot:"Action",overridesResolver:(e,t)=>t.action})({display:"flex",alignItems:"flex-start",padding:"4px 0 0 16px",marginLeft:"auto",marginRight:-8}),Uh={success:m.jsx(X3,{fontSize:"inherit"}),warning:m.jsx(J3,{fontSize:"inherit"}),error:m.jsx(Q3,{fontSize:"inherit"}),info:m.jsx(Y3,{fontSize:"inherit"})},rI=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiAlert"}),{action:s,children:r,className:n,closeText:o="Close",color:l,components:d={},componentsProps:c={},icon:p,iconMapping:u=Uh,onClose:v,role:f="alert",severity:h="success",slotProps:T={},slots:y={},variant:g="standard"}=i,_=U(i,eI),w=q({},i,{color:l,severity:h,variant:g,colorSeverity:l||h}),D=tI(w),x={slots:q({closeButton:d.CloseButton,closeIcon:d.CloseIcon},y),slotProps:q({},c,T)},[k,S]=Wh("closeButton",{elementType:G3,externalForwardedProps:x,ownerState:w}),[A,P]=Wh("closeIcon",{elementType:Z3,externalForwardedProps:x,ownerState:w});return m.jsxs(aI,q({role:f,elevation:0,ownerState:w,className:H(D.root,n),ref:a},_,{children:[p!==!1?m.jsx(iI,{ownerState:w,className:D.icon,children:p||u[h]||Uh[h]}):null,m.jsx(sI,{ownerState:w,className:D.message,children:r}),s!=null?m.jsx($h,{ownerState:w,className:D.action,children:s}):null,s==null&&v?m.jsx($h,{ownerState:w,className:D.action,children:m.jsx(k,q({size:"small","aria-label":o,title:o,color:"inherit",onClick:v},S,{children:m.jsx(A,q({fontSize:"small"},P))}))}):null]}))}),Nn=rI;var Yp={},Zp={};(function(e){Object.defineProperty(e,"__esModule",{value:!0}),e.Doctype=e.CDATA=e.Tag=e.Style=e.Script=e.Comment=e.Directive=e.Text=e.Root=e.isTag=e.ElementType=void 0;var t;(function(i){i.Root="root",i.Text="text",i.Directive="directive",i.Comment="comment",i.Script="script",i.Style="style",i.Tag="tag",i.CDATA="cdata",i.Doctype="doctype"})(t=e.ElementType||(e.ElementType={}));function a(i){return i.type===t.Tag||i.type===t.Script||i.type===t.Style}e.isTag=a,e.Root=t.Root,e.Text=t.Text,e.Directive=t.Directive,e.Comment=t.Comment,e.Script=t.Script,e.Style=t.Style,e.Tag=t.Tag,e.CDATA=t.CDATA,e.Doctype=t.Doctype})(Zp);var Q={},hi=ft&&ft.__extends||function(){var e=function(t,a){return e=Object.setPrototypeOf||{__proto__:[]}instanceof Array&&function(i,s){i.__proto__=s}||function(i,s){for(var r in s)Object.prototype.hasOwnProperty.call(s,r)&&(i[r]=s[r])},e(t,a)};return function(t,a){if(typeof a!="function"&&a!==null)throw new TypeError("Class extends value "+String(a)+" is not a constructor or null");e(t,a);function i(){this.constructor=t}t.prototype=a===null?Object.create(a):(i.prototype=a.prototype,new i)}}(),Tr=ft&&ft.__assign||function(){return Tr=Object.assign||function(e){for(var t,a=1,i=arguments.length;a0?this.children[this.children.length-1]:null},enumerable:!1,configurable:!0}),Object.defineProperty(t.prototype,"childNodes",{get:function(){return this.children},set:function(a){this.children=a},enumerable:!1,configurable:!0}),t}(ef);Q.NodeWithChildren=Gl;var Nb=function(e){hi(t,e);function t(){var a=e!==null&&e.apply(this,arguments)||this;return a.type=vt.ElementType.CDATA,a}return Object.defineProperty(t.prototype,"nodeType",{get:function(){return 4},enumerable:!1,configurable:!0}),t}(Gl);Q.CDATA=Nb;var jb=function(e){hi(t,e);function t(){var a=e!==null&&e.apply(this,arguments)||this;return a.type=vt.ElementType.Root,a}return Object.defineProperty(t.prototype,"nodeType",{get:function(){return 9},enumerable:!1,configurable:!0}),t}(Gl);Q.Document=jb;var Eb=function(e){hi(t,e);function t(a,i,s,r){s===void 0&&(s=[]),r===void 0&&(r=a==="script"?vt.ElementType.Script:a==="style"?vt.ElementType.Style:vt.ElementType.Tag);var n=e.call(this,s)||this;return n.name=a,n.attribs=i,n.type=r,n}return Object.defineProperty(t.prototype,"nodeType",{get:function(){return 1},enumerable:!1,configurable:!0}),Object.defineProperty(t.prototype,"tagName",{get:function(){return this.name},set:function(a){this.name=a},enumerable:!1,configurable:!0}),Object.defineProperty(t.prototype,"attributes",{get:function(){var a=this;return Object.keys(this.attribs).map(function(i){var s,r;return{name:i,value:a.attribs[i],namespace:(s=a["x-attribsNamespace"])===null||s===void 0?void 0:s[i],prefix:(r=a["x-attribsPrefix"])===null||r===void 0?void 0:r[i]}})},enumerable:!1,configurable:!0}),t}(Gl);Q.Element=Eb;function Fb(e){return(0,vt.isTag)(e)}Q.isTag=Fb;function Bb(e){return e.type===vt.ElementType.CDATA}Q.isCDATA=Bb;function Ob(e){return e.type===vt.ElementType.Text}Q.isText=Ob;function Mb(e){return e.type===vt.ElementType.Comment}Q.isComment=Mb;function Lb(e){return e.type===vt.ElementType.Directive}Q.isDirective=Lb;function Wb(e){return e.type===vt.ElementType.Root}Q.isDocument=Wb;function nI(e){return Object.prototype.hasOwnProperty.call(e,"children")}Q.hasChildren=nI;function tf(e,t){t===void 0&&(t=!1);var a;if(Ob(e))a=new Ab(e.data);else if(Mb(e))a=new Rb(e.data);else if(Fb(e)){var i=t?Kd(e.children):[],s=new Eb(e.name,Tr({},e.attribs),i);i.forEach(function(l){return l.parent=s}),e.namespace!=null&&(s.namespace=e.namespace),e["x-attribsNamespace"]&&(s["x-attribsNamespace"]=Tr({},e["x-attribsNamespace"])),e["x-attribsPrefix"]&&(s["x-attribsPrefix"]=Tr({},e["x-attribsPrefix"])),a=s}else if(Bb(e)){var i=t?Kd(e.children):[],r=new Nb(i);i.forEach(function(d){return d.parent=r}),a=r}else if(Wb(e)){var i=t?Kd(e.children):[],n=new jb(i);i.forEach(function(d){return d.parent=n}),e["x-mode"]&&(n["x-mode"]=e["x-mode"]),a=n}else if(Lb(e)){var o=new Ib(e.name,e.data);e["x-name"]!=null&&(o["x-name"]=e["x-name"],o["x-publicId"]=e["x-publicId"],o["x-systemId"]=e["x-systemId"]),a=o}else throw new Error("Not implemented yet: ".concat(e.type));return a.startIndex=e.startIndex,a.endIndex=e.endIndex,e.sourceCodeLocation!=null&&(a.sourceCodeLocation=e.sourceCodeLocation),a}Q.cloneNode=tf;function Kd(e){for(var t=e.map(function(i){return tf(i,!0)}),a=1;a/i,Gh=//i,Lo=function(e,t){throw new Error("This browser does not support `document.implementation.createHTMLDocument`")},mc=function(e,t){throw new Error("This browser does not support `DOMParser.prototype.parseFromString`")},Xh=typeof window=="object"&&window.DOMParser;if(typeof Xh=="function"){var lI=new Xh,dI="text/html";mc=function(e,t){return t&&(e="<".concat(t,">").concat(e,"")),lI.parseFromString(e,dI)},Lo=mc}if(typeof document=="object"&&document.implementation){var En=document.implementation.createHTMLDocument();Lo=function(e,t){if(t){var a=En.documentElement.querySelector(t);return a&&(a.innerHTML=e),En}return En.documentElement.innerHTML=e,En}}var Fn=typeof document=="object"&&document.createElement("template"),hc;Fn&&Fn.content&&(hc=function(e){return Fn.innerHTML=e,Fn.content.childNodes});function uI(e){var t,a,i=e.match(oI),s=i&&i[1]?i[1].toLowerCase():"";switch(s){case Kh:{var r=mc(e);if(!Hh.test(e)){var n=r.querySelector(Vh);(t=n==null?void 0:n.parentNode)===null||t===void 0||t.removeChild(n)}if(!Gh.test(e)){var n=r.querySelector(jn);(a=n==null?void 0:n.parentNode)===null||a===void 0||a.removeChild(n)}return r.querySelectorAll(Kh)}case Vh:case jn:{var o=Lo(e).querySelectorAll(s);return Gh.test(e)&&Hh.test(e)?o[0].parentNode.childNodes:o}default:{if(hc)return hc(e);var n=Lo(e,jn).querySelector(jn);return n.childNodes}}}sf.default=uI;var Rs={},zb={};(function(e){Object.defineProperty(e,"__esModule",{value:!0}),e.CASE_SENSITIVE_TAG_NAMES_MAP=e.CASE_SENSITIVE_TAG_NAMES=void 0,e.CASE_SENSITIVE_TAG_NAMES=["animateMotion","animateTransform","clipPath","feBlend","feColorMatrix","feComponentTransfer","feComposite","feConvolveMatrix","feDiffuseLighting","feDisplacementMap","feDropShadow","feFlood","feFuncA","feFuncB","feFuncG","feFuncR","feGaussianBlur","feImage","feMerge","feMergeNode","feMorphology","feOffset","fePointLight","feSpecularLighting","feSpotLight","feTile","feTurbulence","foreignObject","linearGradient","radialGradient","textPath"],e.CASE_SENSITIVE_TAG_NAMES_MAP=e.CASE_SENSITIVE_TAG_NAMES.reduce(function(t,a){return t[a.toLowerCase()]=a,t},{})})(zb);Object.defineProperty(Rs,"__esModule",{value:!0});Rs.formatDOM=Rs.formatAttributes=void 0;var Bn=Yp,cI=zb;function pI(e){return cI.CASE_SENSITIVE_TAG_NAMES_MAP[e]}function $b(e){for(var t={},a=0,i=e.length;a/;function _I(e){if(typeof e!="string")throw new TypeError("First argument must be a string");if(!e)return[];var t=e.match(yI),a=t?t[1]:void 0;return(0,gI.formatDOM)((0,hI.default)(e),null,a)}af.default=_I;var sa={},Xl={},vI=0;Xl.SAME=vI;var bI=1;Xl.CAMELCASE=bI;Xl.possibleStandardNames={accept:0,acceptCharset:1,"accept-charset":"acceptCharset",accessKey:1,action:0,allowFullScreen:1,alt:0,as:0,async:0,autoCapitalize:1,autoComplete:1,autoCorrect:1,autoFocus:1,autoPlay:1,autoSave:1,capture:0,cellPadding:1,cellSpacing:1,challenge:0,charSet:1,checked:0,children:0,cite:0,class:"className",classID:1,className:1,cols:0,colSpan:1,content:0,contentEditable:1,contextMenu:1,controls:0,controlsList:1,coords:0,crossOrigin:1,dangerouslySetInnerHTML:1,data:0,dateTime:1,default:0,defaultChecked:1,defaultValue:1,defer:0,dir:0,disabled:0,disablePictureInPicture:1,disableRemotePlayback:1,download:0,draggable:0,encType:1,enterKeyHint:1,for:"htmlFor",form:0,formMethod:1,formAction:1,formEncType:1,formNoValidate:1,formTarget:1,frameBorder:1,headers:0,height:0,hidden:0,high:0,href:0,hrefLang:1,htmlFor:1,httpEquiv:1,"http-equiv":"httpEquiv",icon:0,id:0,innerHTML:1,inputMode:1,integrity:0,is:0,itemID:1,itemProp:1,itemRef:1,itemScope:1,itemType:1,keyParams:1,keyType:1,kind:0,label:0,lang:0,list:0,loop:0,low:0,manifest:0,marginWidth:1,marginHeight:1,max:0,maxLength:1,media:0,mediaGroup:1,method:0,min:0,minLength:1,multiple:0,muted:0,name:0,noModule:1,nonce:0,noValidate:1,open:0,optimum:0,pattern:0,placeholder:0,playsInline:1,poster:0,preload:0,profile:0,radioGroup:1,readOnly:1,referrerPolicy:1,rel:0,required:0,reversed:0,role:0,rows:0,rowSpan:1,sandbox:0,scope:0,scoped:0,scrolling:0,seamless:0,selected:0,shape:0,size:0,sizes:0,span:0,spellCheck:1,src:0,srcDoc:1,srcLang:1,srcSet:1,start:0,step:0,style:0,summary:0,tabIndex:1,target:0,title:0,type:0,useMap:1,value:0,width:0,wmode:0,wrap:0,about:0,accentHeight:1,"accent-height":"accentHeight",accumulate:0,additive:0,alignmentBaseline:1,"alignment-baseline":"alignmentBaseline",allowReorder:1,alphabetic:0,amplitude:0,arabicForm:1,"arabic-form":"arabicForm",ascent:0,attributeName:1,attributeType:1,autoReverse:1,azimuth:0,baseFrequency:1,baselineShift:1,"baseline-shift":"baselineShift",baseProfile:1,bbox:0,begin:0,bias:0,by:0,calcMode:1,capHeight:1,"cap-height":"capHeight",clip:0,clipPath:1,"clip-path":"clipPath",clipPathUnits:1,clipRule:1,"clip-rule":"clipRule",color:0,colorInterpolation:1,"color-interpolation":"colorInterpolation",colorInterpolationFilters:1,"color-interpolation-filters":"colorInterpolationFilters",colorProfile:1,"color-profile":"colorProfile",colorRendering:1,"color-rendering":"colorRendering",contentScriptType:1,contentStyleType:1,cursor:0,cx:0,cy:0,d:0,datatype:0,decelerate:0,descent:0,diffuseConstant:1,direction:0,display:0,divisor:0,dominantBaseline:1,"dominant-baseline":"dominantBaseline",dur:0,dx:0,dy:0,edgeMode:1,elevation:0,enableBackground:1,"enable-background":"enableBackground",end:0,exponent:0,externalResourcesRequired:1,fill:0,fillOpacity:1,"fill-opacity":"fillOpacity",fillRule:1,"fill-rule":"fillRule",filter:0,filterRes:1,filterUnits:1,floodOpacity:1,"flood-opacity":"floodOpacity",floodColor:1,"flood-color":"floodColor",focusable:0,fontFamily:1,"font-family":"fontFamily",fontSize:1,"font-size":"fontSize",fontSizeAdjust:1,"font-size-adjust":"fontSizeAdjust",fontStretch:1,"font-stretch":"fontStretch",fontStyle:1,"font-style":"fontStyle",fontVariant:1,"font-variant":"fontVariant",fontWeight:1,"font-weight":"fontWeight",format:0,from:0,fx:0,fy:0,g1:0,g2:0,glyphName:1,"glyph-name":"glyphName",glyphOrientationHorizontal:1,"glyph-orientation-horizontal":"glyphOrientationHorizontal",glyphOrientationVertical:1,"glyph-orientation-vertical":"glyphOrientationVertical",glyphRef:1,gradientTransform:1,gradientUnits:1,hanging:0,horizAdvX:1,"horiz-adv-x":"horizAdvX",horizOriginX:1,"horiz-origin-x":"horizOriginX",ideographic:0,imageRendering:1,"image-rendering":"imageRendering",in2:0,in:0,inlist:0,intercept:0,k1:0,k2:0,k3:0,k4:0,k:0,kernelMatrix:1,kernelUnitLength:1,kerning:0,keyPoints:1,keySplines:1,keyTimes:1,lengthAdjust:1,letterSpacing:1,"letter-spacing":"letterSpacing",lightingColor:1,"lighting-color":"lightingColor",limitingConeAngle:1,local:0,markerEnd:1,"marker-end":"markerEnd",markerHeight:1,markerMid:1,"marker-mid":"markerMid",markerStart:1,"marker-start":"markerStart",markerUnits:1,markerWidth:1,mask:0,maskContentUnits:1,maskUnits:1,mathematical:0,mode:0,numOctaves:1,offset:0,opacity:0,operator:0,order:0,orient:0,orientation:0,origin:0,overflow:0,overlinePosition:1,"overline-position":"overlinePosition",overlineThickness:1,"overline-thickness":"overlineThickness",paintOrder:1,"paint-order":"paintOrder",panose1:0,"panose-1":"panose1",pathLength:1,patternContentUnits:1,patternTransform:1,patternUnits:1,pointerEvents:1,"pointer-events":"pointerEvents",points:0,pointsAtX:1,pointsAtY:1,pointsAtZ:1,prefix:0,preserveAlpha:1,preserveAspectRatio:1,primitiveUnits:1,property:0,r:0,radius:0,refX:1,refY:1,renderingIntent:1,"rendering-intent":"renderingIntent",repeatCount:1,repeatDur:1,requiredExtensions:1,requiredFeatures:1,resource:0,restart:0,result:0,results:0,rotate:0,rx:0,ry:0,scale:0,security:0,seed:0,shapeRendering:1,"shape-rendering":"shapeRendering",slope:0,spacing:0,specularConstant:1,specularExponent:1,speed:0,spreadMethod:1,startOffset:1,stdDeviation:1,stemh:0,stemv:0,stitchTiles:1,stopColor:1,"stop-color":"stopColor",stopOpacity:1,"stop-opacity":"stopOpacity",strikethroughPosition:1,"strikethrough-position":"strikethroughPosition",strikethroughThickness:1,"strikethrough-thickness":"strikethroughThickness",string:0,stroke:0,strokeDasharray:1,"stroke-dasharray":"strokeDasharray",strokeDashoffset:1,"stroke-dashoffset":"strokeDashoffset",strokeLinecap:1,"stroke-linecap":"strokeLinecap",strokeLinejoin:1,"stroke-linejoin":"strokeLinejoin",strokeMiterlimit:1,"stroke-miterlimit":"strokeMiterlimit",strokeWidth:1,"stroke-width":"strokeWidth",strokeOpacity:1,"stroke-opacity":"strokeOpacity",suppressContentEditableWarning:1,suppressHydrationWarning:1,surfaceScale:1,systemLanguage:1,tableValues:1,targetX:1,targetY:1,textAnchor:1,"text-anchor":"textAnchor",textDecoration:1,"text-decoration":"textDecoration",textLength:1,textRendering:1,"text-rendering":"textRendering",to:0,transform:0,typeof:0,u1:0,u2:0,underlinePosition:1,"underline-position":"underlinePosition",underlineThickness:1,"underline-thickness":"underlineThickness",unicode:0,unicodeBidi:1,"unicode-bidi":"unicodeBidi",unicodeRange:1,"unicode-range":"unicodeRange",unitsPerEm:1,"units-per-em":"unitsPerEm",unselectable:0,vAlphabetic:1,"v-alphabetic":"vAlphabetic",values:0,vectorEffect:1,"vector-effect":"vectorEffect",version:0,vertAdvY:1,"vert-adv-y":"vertAdvY",vertOriginX:1,"vert-origin-x":"vertOriginX",vertOriginY:1,"vert-origin-y":"vertOriginY",vHanging:1,"v-hanging":"vHanging",vIdeographic:1,"v-ideographic":"vIdeographic",viewBox:1,viewTarget:1,visibility:0,vMathematical:1,"v-mathematical":"vMathematical",vocab:0,widths:0,wordSpacing:1,"word-spacing":"wordSpacing",writingMode:1,"writing-mode":"writingMode",x1:0,x2:0,x:0,xChannelSelector:1,xHeight:1,"x-height":"xHeight",xlinkActuate:1,"xlink:actuate":"xlinkActuate",xlinkArcrole:1,"xlink:arcrole":"xlinkArcrole",xlinkHref:1,"xlink:href":"xlinkHref",xlinkRole:1,"xlink:role":"xlinkRole",xlinkShow:1,"xlink:show":"xlinkShow",xlinkTitle:1,"xlink:title":"xlinkTitle",xlinkType:1,"xlink:type":"xlinkType",xmlBase:1,"xml:base":"xmlBase",xmlLang:1,"xml:lang":"xmlLang",xmlns:0,"xml:space":"xmlSpace",xmlnsXlink:1,"xmlns:xlink":"xmlnsXlink",xmlSpace:1,y1:0,y2:0,y:0,yChannelSelector:1,z:0,zoomAndPan:1};const Kb=0,gi=1,Jl=2,Ql=3,rf=4,Vb=5,Hb=6;function wI(e){return Ze.hasOwnProperty(e)?Ze[e]:null}function dt(e,t,a,i,s,r,n){this.acceptsBooleans=t===Jl||t===Ql||t===rf,this.attributeName=i,this.attributeNamespace=s,this.mustUseProperty=a,this.propertyName=e,this.type=t,this.sanitizeURL=r,this.removeEmptyString=n}const Ze={},TI=["children","dangerouslySetInnerHTML","defaultValue","defaultChecked","innerHTML","suppressContentEditableWarning","suppressHydrationWarning","style"];TI.forEach(e=>{Ze[e]=new dt(e,Kb,!1,e,null,!1,!1)});[["acceptCharset","accept-charset"],["className","class"],["htmlFor","for"],["httpEquiv","http-equiv"]].forEach(([e,t])=>{Ze[e]=new dt(e,gi,!1,t,null,!1,!1)});["contentEditable","draggable","spellCheck","value"].forEach(e=>{Ze[e]=new dt(e,Jl,!1,e.toLowerCase(),null,!1,!1)});["autoReverse","externalResourcesRequired","focusable","preserveAlpha"].forEach(e=>{Ze[e]=new dt(e,Jl,!1,e,null,!1,!1)});["allowFullScreen","async","autoFocus","autoPlay","controls","default","defer","disabled","disablePictureInPicture","disableRemotePlayback","formNoValidate","hidden","loop","noModule","noValidate","open","playsInline","readOnly","required","reversed","scoped","seamless","itemScope"].forEach(e=>{Ze[e]=new dt(e,Ql,!1,e.toLowerCase(),null,!1,!1)});["checked","multiple","muted","selected"].forEach(e=>{Ze[e]=new dt(e,Ql,!0,e,null,!1,!1)});["capture","download"].forEach(e=>{Ze[e]=new dt(e,rf,!1,e,null,!1,!1)});["cols","rows","size","span"].forEach(e=>{Ze[e]=new dt(e,Hb,!1,e,null,!1,!1)});["rowSpan","start"].forEach(e=>{Ze[e]=new dt(e,Vb,!1,e.toLowerCase(),null,!1,!1)});const nf=/[\-\:]([a-z])/g,of=e=>e[1].toUpperCase();["accent-height","alignment-baseline","arabic-form","baseline-shift","cap-height","clip-path","clip-rule","color-interpolation","color-interpolation-filters","color-profile","color-rendering","dominant-baseline","enable-background","fill-opacity","fill-rule","flood-color","flood-opacity","font-family","font-size","font-size-adjust","font-stretch","font-style","font-variant","font-weight","glyph-name","glyph-orientation-horizontal","glyph-orientation-vertical","horiz-adv-x","horiz-origin-x","image-rendering","letter-spacing","lighting-color","marker-end","marker-mid","marker-start","overline-position","overline-thickness","paint-order","panose-1","pointer-events","rendering-intent","shape-rendering","stop-color","stop-opacity","strikethrough-position","strikethrough-thickness","stroke-dasharray","stroke-dashoffset","stroke-linecap","stroke-linejoin","stroke-miterlimit","stroke-opacity","stroke-width","text-anchor","text-decoration","text-rendering","underline-position","underline-thickness","unicode-bidi","unicode-range","units-per-em","v-alphabetic","v-hanging","v-ideographic","v-mathematical","vector-effect","vert-adv-y","vert-origin-x","vert-origin-y","word-spacing","writing-mode","xmlns:xlink","x-height"].forEach(e=>{const t=e.replace(nf,of);Ze[t]=new dt(t,gi,!1,e,null,!1,!1)});["xlink:actuate","xlink:arcrole","xlink:role","xlink:show","xlink:title","xlink:type"].forEach(e=>{const t=e.replace(nf,of);Ze[t]=new dt(t,gi,!1,e,"http://www.w3.org/1999/xlink",!1,!1)});["xml:base","xml:lang","xml:space"].forEach(e=>{const t=e.replace(nf,of);Ze[t]=new dt(t,gi,!1,e,"http://www.w3.org/XML/1998/namespace",!1,!1)});["tabIndex","crossOrigin"].forEach(e=>{Ze[e]=new dt(e,gi,!1,e.toLowerCase(),null,!1,!1)});const kI="xlinkHref";Ze[kI]=new dt("xlinkHref",gi,!1,"xlink:href","http://www.w3.org/1999/xlink",!0,!1);["src","href","action","formAction"].forEach(e=>{Ze[e]=new dt(e,gi,!1,e.toLowerCase(),null,!0,!0)});const{CAMELCASE:qI,SAME:DI,possibleStandardNames:Jh}=Xl,xI=":A-Z_a-z\\u00C0-\\u00D6\\u00D8-\\u00F6\\u00F8-\\u02FF\\u0370-\\u037D\\u037F-\\u1FFF\\u200C-\\u200D\\u2070-\\u218F\\u2C00-\\u2FEF\\u3001-\\uD7FF\\uF900-\\uFDCF\\uFDF0-\\uFFFD",CI=xI+"\\-.0-9\\u00B7\\u0300-\\u036F\\u203F-\\u2040",SI=RegExp.prototype.test.bind(new RegExp("^(data|aria)-["+CI+"]*$")),PI=Object.keys(Jh).reduce((e,t)=>{const a=Jh[t];return a===DI?e[t]=t:a===qI?e[t.toLowerCase()]=t:e[t]=a,e},{});sa.BOOLEAN=Ql;sa.BOOLEANISH_STRING=Jl;sa.NUMERIC=Vb;sa.OVERLOADED_BOOLEAN=rf;sa.POSITIVE_NUMERIC=Hb;sa.RESERVED=Kb;sa.STRING=gi;sa.getPropertyInfo=wI;sa.isCustomAttribute=SI;sa.possibleStandardNames=PI;var lf={},df={},Qh=/\/\*[^*]*\*+([^/*][^*]*\*+)*\//g,AI=/\n/g,RI=/^\s*/,II=/^(\*?[-#/*\\\w]+(\[[0-9a-z_-]+\])?)\s*/,NI=/^:\s*/,jI=/^((?:'(?:\\'|.)*?'|"(?:\\"|.)*?"|\([^)]*?\)|[^};])+)/,EI=/^[;\s]*/,FI=/^\s+|\s+$/g,BI=` +`,Yh="/",Zh="*",qi="",OI="comment",MI="declaration",LI=function(e,t){if(typeof e!="string")throw new TypeError("First argument must be a string");if(!e)return[];t=t||{};var a=1,i=1;function s(f){var h=f.match(AI);h&&(a+=h.length);var T=f.lastIndexOf(BI);i=~T?f.length-T:i+f.length}function r(){var f={line:a,column:i};return function(h){return h.position=new n(f),d(),h}}function n(f){this.start=f,this.end={line:a,column:i},this.source=t.source}n.prototype.content=e;function o(f){var h=new Error(t.source+":"+a+":"+i+": "+f);if(h.reason=f,h.filename=t.source,h.line=a,h.column=i,h.source=e,!t.silent)throw h}function l(f){var h=f.exec(e);if(h){var T=h[0];return s(T),e=e.slice(T.length),h}}function d(){l(RI)}function c(f){var h;for(f=f||[];h=p();)h!==!1&&f.push(h);return f}function p(){var f=r();if(!(Yh!=e.charAt(0)||Zh!=e.charAt(1))){for(var h=2;qi!=e.charAt(h)&&(Zh!=e.charAt(h)||Yh!=e.charAt(h+1));)++h;if(h+=2,qi===e.charAt(h-1))return o("End of comment missing");var T=e.slice(2,h-2);return i+=2,s(T),e=e.slice(h),i+=2,f({type:OI,comment:T})}}function u(){var f=r(),h=l(II);if(h){if(p(),!l(NI))return o("property missing ':'");var T=l(jI),y=f({type:MI,property:eg(h[0].replace(Qh,qi)),value:T?eg(T[0].replace(Qh,qi)):qi});return l(EI),y}}function v(){var f=[];c(f);for(var h;h=u();)h!==!1&&(f.push(h),c(f));return f}return d(),v()};function eg(e){return e?e.replace(FI,qi):qi}var WI=ft&&ft.__importDefault||function(e){return e&&e.__esModule?e:{default:e}};Object.defineProperty(df,"__esModule",{value:!0});var zI=WI(LI);function $I(e,t){var a=null;if(!e||typeof e!="string")return a;var i=(0,zI.default)(e),s=typeof t=="function";return i.forEach(function(r){if(r.type==="declaration"){var n=r.property,o=r.value;s?t(n,o,r):o&&(a=a||{},a[n]=o)}}),a}df.default=$I;var Yl={};Object.defineProperty(Yl,"__esModule",{value:!0});Yl.camelCase=void 0;var UI=/^--[a-zA-Z0-9-]+$/,KI=/-([a-z])/g,VI=/^[^-]+$/,HI=/^-(webkit|moz|ms|o|khtml)-/,GI=/^-(ms)-/,XI=function(e){return!e||VI.test(e)||UI.test(e)},JI=function(e,t){return t.toUpperCase()},tg=function(e,t){return"".concat(t,"-")},QI=function(e,t){return t===void 0&&(t={}),XI(e)?e:(e=e.toLowerCase(),t.reactCompat?e=e.replace(GI,tg):e=e.replace(HI,tg),e.replace(KI,JI))};Yl.camelCase=QI;var YI=ft&&ft.__importDefault||function(e){return e&&e.__esModule?e:{default:e}};Object.defineProperty(lf,"__esModule",{value:!0});var ZI=YI(df),eN=Yl;function tN(e,t){var a={};return!e||typeof e!="string"||(0,ZI.default)(e,function(i,s){i&&s&&(a[(0,eN.camelCase)(i,t)]=s)}),a}lf.default=tN;var aN=b,iN=lf.default,sN=new Set(["annotation-xml","color-profile","font-face","font-face-src","font-face-uri","font-face-format","font-face-name","missing-glyph"]);function rN(e,t){return e.indexOf("-")===-1?t&&typeof t.is=="string":!sN.has(e)}var nN={reactCompat:!0};function oN(e,t){if(e!=null)try{t.style=iN(e,nN)}catch{t.style={}}}var lN=aN.version.split(".")[0]>=16,Gb=new Set(["tr","tbody","thead","tfoot","colgroup","table","head","html","frameset"]);function dN(e){return!Gb.has(e.name)}function uN(e){return e}var Xb={PRESERVE_CUSTOM_ATTRIBUTES:lN,ELEMENTS_WITH_NO_TEXT_CHILDREN:Gb,isCustomComponent:rN,setStyleProp:oN,canTextBeChildOfNode:dN,returnFirstArg:uN},dr=sa,ag=Xb,cN=["checked","value"],pN=["input","select","textarea"],fN={reset:!0,submit:!0},Jb=function(t,a){t=t||{};var i,s,r,n,o,l={},d=t.type&&fN[t.type];for(i in t){if(r=t[i],dr.isCustomAttribute(i)){l[i]=r;continue}if(s=i.toLowerCase(),n=ig(s),n){switch(o=dr.getPropertyInfo(n),cN.indexOf(n)!==-1&&pN.indexOf(a)!==-1&&!d&&(n=ig("default"+s)),l[n]=r,o&&o.type){case dr.BOOLEAN:l[n]=!0;break;case dr.OVERLOADED_BOOLEAN:r===""&&(l[n]=!0);break}continue}ag.PRESERVE_CUSTOM_ATTRIBUTES&&(l[i]=r)}return ag.setStyleProp(t.style,l),l};function ig(e){return dr.possibleStandardNames[e]}var mN=b,hN=Jb,Vr=Xb,gN=Vr.setStyleProp,yN=Vr.canTextBeChildOfNode;function Qb(e,t){t=t||{};for(var a=t.library||mN,i=a.cloneElement,s=a.createElement,r=a.isValidElement,n=[],o,l,d=typeof t.replace=="function",c=t.transform||Vr.returnFirstArg,p,u,v,f=t.trim,h=0,T=e.length;h1&&(p=i(p,{key:p.key||h})),n.push(c(p,o,h));continue}if(o.type==="text"){if(l=!o.data.trim().length,l&&o.parent&&!yN(o.parent)||f&&l)continue;n.push(c(o.data,o,h));continue}switch(u=o.attribs,_N(o)?gN(u.style,u):u&&(u=hN(u,o.name)),v=null,o.type){case"script":case"style":o.children[0]&&(u.dangerouslySetInnerHTML={__html:o.children[0].data});break;case"tag":o.name==="textarea"&&o.children[0]?u.defaultValue=o.children[0].data:o.children&&o.children.length&&(v=Qb(o.children,t));break;default:continue}T>1&&(u.key=h),n.push(c(s(o.name,u,v),o,h))}return n.length===1?n[0]:n}function _N(e){return Vr.PRESERVE_CUSTOM_ATTRIBUTES&&e.type==="tag"&&Vr.isCustomComponent(e.name,e.attribs)}var vN=Qb,Zl=Yp,cs=af.default,bN=Jb,Yb=vN;cs=typeof cs.default=="function"?cs.default:cs;var wN={lowerCaseAttributeNames:!1};function ga(e,t){if(typeof e!="string")throw new TypeError("First argument must be a string");return e===""?[]:(t=t||{},Yb(cs(e,t.htmlparser2||wN),t))}ga.domToReact=Yb;ga.htmlToDOM=cs;ga.attributesToProps=bN;ga.Comment=Zl.Comment;ga.Element=Zl.Element;ga.ProcessingInstruction=Zl.ProcessingInstruction;ga.Text=Zl.Text;var TN=ga;ga.default=ga;const Ea=gc(TN);Ea.domToReact;Ea.htmlToDOM;Ea.attributesToProps;Ea.Comment;Ea.Element;Ea.ProcessingInstruction;Ea.Text;const sg=ya.entrypointtypes,kN=ya.plugins,rg=ya.status_dict,qN={}.VITE_PR_PREVIEW_PATH||"/aiida-registry/";function DN({pluginKey:e}){const t=kN[e];return b.useEffect(()=>{window.scrollTo(0,0),document.documentElement.style.scrollBehavior="smooth",(()=>{const i=window.location.hash;if(i){let s=window.location.href;window.location.href=s+" ",window.location.href=s;const r=document.getElementById(i);r&&r.scrollIntoView()}})()},[]),m.jsx(m.Fragment,{children:m.jsxs("div",{id:"details",className:"fade-enter",children:[m.jsxs("h1",{className:"plugin-header",children:['AiiDA plugin package "',m.jsx("a",{href:t.code_home,children:t.name}),'"']}),m.jsx(Bi,{to:"/",children:m.jsx("p",{style:{display:"inline"},children:"< back to the registry index"})}),m.jsx("h2",{id:"general.information",children:"General information"}),m.jsxs("div",{children:[m.jsxs("p",{children:[m.jsx("strong",{children:"Current state: "}),m.jsx("img",{className:"svg-badge",src:`${qN}${rg[t.development_status][1]}`,title:rg[t.development_status][0]})]}),t.metadata.description&&m.jsxs("p",{children:[m.jsx("strong",{children:"Short description"}),": ",t.metadata.description]}),t.pip_url&&m.jsxs("p",{children:[m.jsx("strong",{children:"How to install"}),": ",m.jsx("code",{children:t.pip_install_cmd})]}),m.jsxs("p",{children:[m.jsx("strong",{children:"Source code"}),": ",m.jsx("a",{href:t.code_home,target:"_blank",children:"Go to the source code repository"})]}),t.documentation_url?m.jsxs("p",{children:[m.jsx("strong",{children:"Documentation"}),": ",m.jsx("a",{href:t.documentation_url,target:"_blank",children:"Go to plugin documentation"})]}):m.jsxs("p",{children:[m.jsx("strong",{children:"Documentation"}),": No documentation provided by the package author"]})]}),m.jsx("h3",{children:"Registry checks"}),t.warnings.length!==0||t.errors.length!==0?m.jsxs(m.Fragment,{children:[t.warnings&&m.jsx(m.Fragment,{children:t.warnings.map(a=>m.jsx(Nn,{severity:"warning",children:Ea(`${a}`)}))}),t.errors&&m.jsx(m.Fragment,{children:t.errors.map(a=>m.jsx(Nn,{severity:"error",children:Ea(`${a}`)}))}),m.jsxs(Nn,{severity:"info",children:["Click the warning/error code will redirect to ",m.jsx("a",{href:"https://github.com/aiidateam/aiida-registry#how-to-fix-registry-warnings-and-errors",children:"troubleshooting section"})," for the fix of the issue."]})]}):m.jsx(Nn,{severity:"success",children:"All checks passed!"}),m.jsx("h2",{id:"detailed.information",children:"Detailed information"}),Object.keys(t.metadata).length!==0?m.jsxs(m.Fragment,{children:[t.metadata.author&&m.jsxs("p",{children:[m.jsx("strong",{children:"Author(s)"}),": ",t.metadata.author]}),t.metadata.author_email&&m.jsxs("p",{children:[m.jsx("strong",{children:"Contact"}),":",t.metadata.author_email.split(",").map(a=>m.jsxs("span",{children:[m.jsx("a",{href:`mailto:${a.trim()}`,children:a.trim()}),", "]},a))]}),m.jsxs("p",{children:[m.jsx("strong",{children:"How to use from python"}),":"," ",m.jsxs("code",{children:["import ",t.package_name]})]}),m.jsxs("p",{children:[m.jsx("strong",{children:"Most recent version"}),": ",t.metadata.version]}),t.aiida_version&&m.jsxs("p",{children:[m.jsx("strong",{children:"Compatibility: "}),m.jsx("img",{className:"svg-badge",src:`https://img.shields.io/badge/AiiDA-${t.aiida_version}-007ec6.svg?logo=${ub}`})]}),t.summaryinfo.length!==0&&m.jsxs(m.Fragment,{children:[m.jsx("h3",{id:"plugins",children:"Plugins provided by the package"}),t.summaryinfo.map(a=>m.jsxs("span",{className:"badge",children:[m.jsx("span",{className:`badge-left ${a.colorclass}`,children:a.text}),m.jsx("span",{className:"badge-right",children:a.count})]},a.text))]}),t.entry_points?Object.entries(t.entry_points).map(([a,i])=>m.jsx(m.Fragment,{children:m.jsxs("div",{children:[m.jsx("h2",{style:{color:"black"},id:a,children:a in sg?m.jsxs(m.Fragment,{children:[sg[a]," ",m.jsxs("span",{className:"entrypointraw",children:["(",a,")"]})]}):a}),m.jsx("ul",{children:Object.entries(i).map(([s,r])=>m.jsxs("li",{children:[m.jsx("h2",{style:{color:"black"},id:`${a}.${s}`,children:s}),typeof r=="string"?m.jsxs("div",{className:"classbox",children:["class",m.jsxs("span",{className:"tooltiptext",children:[" ",r]})]}):m.jsx(xN,{entryPoints:r})]},s))})]},a)})):m.jsx("p",{children:"No entry points defined for this plugin."})]}):m.jsx("div",{id:"description",children:m.jsxs("p",{children:["Detailed information for this package could not be obtained. Ask the plugin author to add a ",m.jsx("code",{children:"setup.json"})," file to the plugin source code."]})})]})})}const xN=({entryPoints:e})=>m.jsxs("div",{style:{overflow:"auto"},children:[m.jsx("table",{children:m.jsx("tbody",{children:m.jsxs("tr",{children:[m.jsx("th",{children:"Class"}),m.jsx("td",{children:m.jsx("code",{children:e.class})})]})})}),m.jsxs("table",{children:[m.jsx("tr",{children:m.jsx("th",{children:"Description"})}),e.description.map(t=>m.jsx("tr",{className:"ep_description",children:m.jsx(fc,{children:t.trim()})}))]}),m.jsxs("table",{children:[m.jsxs("tr",{children:[m.jsx("th",{children:"Inputs"}),m.jsx("th",{children:"Required"}),m.jsx("th",{children:"Valid Types"}),m.jsx("th",{children:"Description"})]}),m.jsx(ng,{spec:e.spec.inputs}),m.jsxs("tr",{children:[m.jsx("th",{children:"Outputs"}),m.jsx("th",{children:"Required"}),m.jsx("th",{children:"Valid Types"}),m.jsx("th",{children:"Description"})]}),m.jsx(ng,{spec:e.spec.outputs})]}),m.jsxs("table",{children:[m.jsx("tr",{children:m.jsx("th",{children:"Exit Codes"})}),m.jsxs("tr",{children:[m.jsx("th",{children:"Status"}),m.jsx("th",{children:"Message"})]}),e.spec.exit_codes.map(t=>m.jsxs("tr",{className:"ep_description",children:[m.jsx("td",{children:t.status}),m.jsx(fc,{children:t.message})]}))]})]}),ng=({spec:e})=>m.jsx(m.Fragment,{children:e.map(t=>m.jsxs("tr",{className:"ep_description",children:[m.jsx("td",{children:t.name}),m.jsx("td",{children:t.required.toString()}),m.jsx("td",{children:t.valid_types}),m.jsx(fc,{children:t.info})]}))});const CN=["addEndListener","appear","children","container","direction","easing","in","onEnter","onEntered","onEntering","onExit","onExited","onExiting","style","timeout","TransitionComponent"];function SN(e,t,a){const i=t.getBoundingClientRect(),s=a&&a.getBoundingClientRect(),r=ha(t);let n;if(t.fakeTransform)n=t.fakeTransform;else{const d=r.getComputedStyle(t);n=d.getPropertyValue("-webkit-transform")||d.getPropertyValue("transform")}let o=0,l=0;if(n&&n!=="none"&&typeof n=="string"){const d=n.split("(")[1].split(")")[0].split(",");o=parseInt(d[4],10),l=parseInt(d[5],10)}return e==="left"?s?`translateX(${s.right+o-i.left}px)`:`translateX(${r.innerWidth+o-i.left}px)`:e==="right"?s?`translateX(-${i.right-s.left-o}px)`:`translateX(-${i.left+i.width-o}px)`:e==="up"?s?`translateY(${s.bottom+l-i.top}px)`:`translateY(${r.innerHeight+l-i.top}px)`:s?`translateY(-${i.top-s.top+i.height-l}px)`:`translateY(-${i.top+i.height-l}px)`}function PN(e){return typeof e=="function"?e():e}function On(e,t,a){const i=PN(a),s=SN(e,t,i);s&&(t.style.webkitTransform=s,t.style.transform=s)}const AN=b.forwardRef(function(t,a){const i=on(),s={enter:i.transitions.easing.easeOut,exit:i.transitions.easing.sharp},r={enter:i.transitions.duration.enteringScreen,exit:i.transitions.duration.leavingScreen},{addEndListener:n,appear:o=!0,children:l,container:d,direction:c="down",easing:p=s,in:u,onEnter:v,onEntered:f,onEntering:h,onExit:T,onExited:y,onExiting:g,style:_,timeout:w=r,TransitionComponent:D=Wp}=t,x=U(t,CN),k=b.useRef(null),S=Ke(l.ref,k,a),A=C=>N=>{C&&(N===void 0?C(k.current):C(k.current,N))},P=A((C,N)=>{On(c,C,d),Up(C),v&&v(C,N)}),j=A((C,N)=>{const L=Ps({timeout:w,style:_,easing:p},{mode:"enter"});C.style.webkitTransition=i.transitions.create("-webkit-transform",q({},L)),C.style.transition=i.transitions.create("transform",q({},L)),C.style.webkitTransform="none",C.style.transform="none",h&&h(C,N)}),E=A(f),F=A(g),I=A(C=>{const N=Ps({timeout:w,style:_,easing:p},{mode:"exit"});C.style.webkitTransition=i.transitions.create("-webkit-transform",N),C.style.transition=i.transitions.create("transform",N),On(c,C,d),T&&T(C)}),B=A(C=>{C.style.webkitTransition="",C.style.transition="",y&&y(C)}),O=C=>{n&&n(k.current,C)},W=b.useCallback(()=>{k.current&&On(c,k.current,d)},[c,d]);return b.useEffect(()=>{if(u||c==="down"||c==="right")return;const C=hl(()=>{k.current&&On(c,k.current,d)}),N=ha(k.current);return N.addEventListener("resize",C),()=>{C.clear(),N.removeEventListener("resize",C)}},[c,u,d]),b.useEffect(()=>{u||W()},[u,W]),m.jsx(D,q({nodeRef:k,onEnter:P,onEntered:E,onEntering:j,onExit:I,onExited:B,onExiting:F,addEndListener:O,appear:o,in:u,timeout:w},x,{children:(C,N)=>b.cloneElement(l,q({ref:S,style:q({visibility:C==="exited"&&!u?"hidden":void 0},_,l.props.style)},N))}))}),RN=AN;function IN(e){return pe("MuiDrawer",e)}re("MuiDrawer",["root","docked","paper","paperAnchorLeft","paperAnchorRight","paperAnchorTop","paperAnchorBottom","paperAnchorDockedLeft","paperAnchorDockedRight","paperAnchorDockedTop","paperAnchorDockedBottom","modal"]);const NN=["BackdropProps"],jN=["anchor","BackdropProps","children","className","elevation","hideBackdrop","ModalProps","onClose","open","PaperProps","SlideProps","TransitionComponent","transitionDuration","variant"],Zb=(e,t)=>{const{ownerState:a}=e;return[t.root,(a.variant==="permanent"||a.variant==="persistent")&&t.docked,t.modal]},EN=e=>{const{classes:t,anchor:a,variant:i}=e,s={root:["root"],docked:[(i==="permanent"||i==="persistent")&&"docked"],modal:["modal"],paper:["paper",`paperAnchor${X(a)}`,i!=="temporary"&&`paperAnchorDocked${X(a)}`]};return fe(s,IN,t)},FN=z(Kp,{name:"MuiDrawer",slot:"Root",overridesResolver:Zb})(({theme:e})=>({zIndex:(e.vars||e).zIndex.drawer})),og=z("div",{shouldForwardProp:It,name:"MuiDrawer",slot:"Docked",skipVariantsResolver:!1,overridesResolver:Zb})({flex:"0 0 auto"}),BN=z(un,{name:"MuiDrawer",slot:"Paper",overridesResolver:(e,t)=>{const{ownerState:a}=e;return[t.paper,t[`paperAnchor${X(a.anchor)}`],a.variant!=="temporary"&&t[`paperAnchorDocked${X(a.anchor)}`]]}})(({theme:e,ownerState:t})=>q({overflowY:"auto",display:"flex",flexDirection:"column",height:"100%",flex:"1 0 auto",zIndex:(e.vars||e).zIndex.drawer,WebkitOverflowScrolling:"touch",position:"fixed",top:0,outline:0},t.anchor==="left"&&{left:0},t.anchor==="top"&&{top:0,left:0,right:0,height:"auto",maxHeight:"100%"},t.anchor==="right"&&{right:0},t.anchor==="bottom"&&{top:"auto",left:0,bottom:0,right:0,height:"auto",maxHeight:"100%"},t.anchor==="left"&&t.variant!=="temporary"&&{borderRight:`1px solid ${(e.vars||e).palette.divider}`},t.anchor==="top"&&t.variant!=="temporary"&&{borderBottom:`1px solid ${(e.vars||e).palette.divider}`},t.anchor==="right"&&t.variant!=="temporary"&&{borderLeft:`1px solid ${(e.vars||e).palette.divider}`},t.anchor==="bottom"&&t.variant!=="temporary"&&{borderTop:`1px solid ${(e.vars||e).palette.divider}`})),e0={left:"right",right:"left",top:"down",bottom:"up"};function ON(e){return["left","right"].indexOf(e)!==-1}function MN({direction:e},t){return e==="rtl"&&ON(t)?e0[t]:t}const LN=b.forwardRef(function(t,a){const i=me({props:t,name:"MuiDrawer"}),s=on(),r=Uv(),n={enter:s.transitions.duration.enteringScreen,exit:s.transitions.duration.leavingScreen},{anchor:o="left",BackdropProps:l,children:d,className:c,elevation:p=16,hideBackdrop:u=!1,ModalProps:{BackdropProps:v}={},onClose:f,open:h=!1,PaperProps:T={},SlideProps:y,TransitionComponent:g=RN,transitionDuration:_=n,variant:w="temporary"}=i,D=U(i.ModalProps,NN),x=U(i,jN),k=b.useRef(!1);b.useEffect(()=>{k.current=!0},[]);const S=MN({direction:r?"rtl":"ltr"},o),P=q({},i,{anchor:o,elevation:p,open:h,variant:w},x),j=EN(P),E=m.jsx(BN,q({elevation:w==="temporary"?p:0,square:!0},T,{className:H(j.paper,T.className),ownerState:P,children:d}));if(w==="permanent")return m.jsx(og,q({className:H(j.root,j.docked,c),ownerState:P,ref:a},x,{children:E}));const F=m.jsx(g,q({in:h,direction:e0[S],timeout:_,appear:k.current},y,{children:E}));return w==="persistent"?m.jsx(og,q({className:H(j.root,j.docked,c),ownerState:P,ref:a},x,{children:F})):m.jsx(FN,q({BackdropProps:q({},l,v,{transitionDuration:_}),className:H(j.root,j.modal,c),open:h,ownerState:P,onClose:f,hideBackdrop:u,ref:a},x,D,{children:F}))}),WN=LN,zN=ya.plugins;function $N({pluginKey:e}){const t=zN[e];function a(){function s(){document.querySelector("header").style.top="-155px",document.querySelector("#sidebar .MuiDrawer-paper").style.marginTop="0"}setTimeout(s,800)}const i=m.jsxs("div",{style:{paddingLeft:"10px"},children:[m.jsx("h1",{children:"Plugin content"}),m.jsx(_h,{}),m.jsx("p",{children:m.jsx("a",{style:{color:"black"},href:"#general.information",onClick:a,children:"General Information"})}),m.jsx("p",{children:m.jsx("a",{style:{color:"black"},href:"#detailed.information",onClick:a,children:"Detailed Information"})}),m.jsx("p",{children:m.jsx("a",{style:{color:"black"},href:"#plugins",onClick:a,children:"Plugins provided by the package"})}),t.entry_points&&Object.entries(t.entry_points).map(([s,r])=>m.jsx(m.Fragment,{children:m.jsx("ul",{children:m.jsxs("li",{children:[m.jsx("a",{style:{color:"black"},href:`#${s}`,onClick:a,children:s}),Object.entries(r).map(([n,o])=>m.jsx("ul",{children:m.jsx("li",{children:m.jsx("a",{style:{color:"black"},href:`#${s}.${n}`,onClick:a,children:n})})},n))]})})})),m.jsx(_h,{})]});return m.jsx(WN,{variant:"permanent",id:"sidebar",anchor:"right",sx:{display:{xs:"none",sm:"block"}},open:!0,children:i})}function UN(){return m.jsxs(m.Fragment,{children:[m.jsx(KN,{}),m.jsx("div",{style:{marginTop:"155px"},children:m.jsx(UA,{children:m.jsx(FC,{children:m.jsxs(vT,{children:[m.jsx(Lu,{path:"/",element:m.jsx(NR,{})}),m.jsx(Lu,{path:"/:key",element:m.jsx(HN,{})})]})})})}),m.jsx(VN,{})]})}function KN(){return m.jsx("header",{children:m.jsxs("div",{style:{paddingLeft:"20px"},children:[m.jsx("h1",{children:m.jsx("a",{href:"http://aiidateam.github.io/aiida-registry",children:m.jsx("img",{src:AT,height:"70px"})})}),m.jsx("p",{style:{fontSize:"90%"},children:m.jsx("a",{href:"http://github.com/aiidateam/aiida-registry",style:{color:"#999"},children:"[View on GitHub/register your package]"})})]})})}function VN(){return m.jsxs("footer",{className:"footer",children:[m.jsx("hr",{}),"The official ",m.jsx("a",{href:"http://aiidateam.github.io/aiida-registry",children:"registry"})," of ",m.jsx("a",{href:"http://www.aiida.net",children:"AiiDA"})," plugins.",m.jsx("br",{}),"This work is supported by the ",m.jsx("a",{href:"http://nccr-marvel.ch",target:"_blank",children:"MARVEL National Centre for Competence in Research"})," funded by the ",m.jsx("a",{href:"http://www.snf.ch/en",target:"_blank",children:"Swiss National Science Foundation"}),", as well as by the ",m.jsx("a",{href:"http://www.max-centre.eu",target:"_blank",children:"MaX European Centre of Excellence"})," funded by the Horizon 2020 EINFRA-5 program, Grant No. 676598.",m.jsx("br",{}),m.jsx("br",{}),m.jsxs("div",{style:{textAlign:"center"},children:[m.jsx("img",{src:RT,height:"70px"}),"    ",m.jsx("img",{src:IT,height:"70px"})]})]})}function HN(){const{key:e}=aT();b.useEffect(()=>(document.querySelector("footer").style.width="calc(100% - 380px)",()=>{document.querySelector("footer").style.width="calc(100% - 64px)"}),[]);function t(){var a=window.scrollY;window.onscroll=function(){var i=window.scrollY;a>i?(document.querySelector("header").style.top="0",document.querySelector("#sidebar .MuiDrawer-paper").style.marginTop="155px"):a>150&&(document.querySelector("header").style.top="-155px",document.querySelector("#sidebar .MuiDrawer-paper").style.marginTop="0"),a=i}}return t(),m.jsx(m.Fragment,{children:m.jsxs("div",{id:"detailsContainer",children:[m.jsx(DN,{pluginKey:e}),m.jsx($N,{pluginKey:e})]})})}const GN={}.VITE_PR_PREVIEW_PATH||"/aiida-registry/";Vd.createRoot(document.getElementById("root")).render(m.jsx(ea.StrictMode,{children:m.jsx(xT,{basename:GN,children:m.jsx(UN,{})})})); diff --git a/index.html b/index.html index 4094e460..413dd9ba 100644 --- a/index.html +++ b/index.html @@ -14,7 +14,7 @@ AiiDA Plugin Registry - + diff --git a/plugins_metadata.json b/plugins_metadata.json index 5640ed41..a8623dd5 100644 --- a/plugins_metadata.json +++ b/plugins_metadata.json @@ -1148,7 +1148,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-alloy
Collecting git+https://github.com/DanielMarchand/aiida-alloy\n  Cloning https://github.com/DanielMarchand/aiida-alloy to /tmp/pip-req-build-x3wq1q60\n  Running command git clone --filter=blob:none --quiet https://github.com/DanielMarchand/aiida-alloy /tmp/pip-req-build-x3wq1q60\n  Resolved https://github.com/DanielMarchand/aiida-alloy to commit dfae54b29291dfe94d4a5ca136738020f447bc1e\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'error'\n  error: subprocess-exited-with-error\n  \n  \u00d7 python setup.py egg_info did not run successfully.\n  \u2502 exit code: 1\n  \u2570\u2500> [6 lines of output]\n      Traceback (most recent call last):\n        File \"\", line 2, in \n        File \"\", line 34, in \n        File \"/tmp/pip-req-build-x3wq1q60/setup.py\", line 20, in \n          long_description=open('README.md').read(),\n      FileNotFoundError: [Errno 2] No such file or directory: 'README.md'\n      [end of output]\n  \n  note: This error originates from a subprocess, and is likely not a problem with pip.\nerror: metadata-generation-failed\n\n\u00d7 Encountered error while generating package metadata.\n\u2570\u2500> See above for output.\n\nnote: This is an issue with the package mentioned above, not pip.\nhint: See above for details.\n
" + "E001: Failed to install plugin aiida-alloy
Collecting git+https://github.com/DanielMarchand/aiida-alloy\n  Cloning https://github.com/DanielMarchand/aiida-alloy to /tmp/pip-req-build-8h1hn0so\n  Running command git clone --filter=blob:none --quiet https://github.com/DanielMarchand/aiida-alloy /tmp/pip-req-build-8h1hn0so\n  Resolved https://github.com/DanielMarchand/aiida-alloy to commit dfae54b29291dfe94d4a5ca136738020f447bc1e\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'error'\n  error: subprocess-exited-with-error\n  \n  \u00d7 python setup.py egg_info did not run successfully.\n  \u2502 exit code: 1\n  \u2570\u2500> [6 lines of output]\n      Traceback (most recent call last):\n        File \"\", line 2, in \n        File \"\", line 34, in \n        File \"/tmp/pip-req-build-8h1hn0so/setup.py\", line 20, in \n          long_description=open('README.md').read(),\n      FileNotFoundError: [Errno 2] No such file or directory: 'README.md'\n      [end of output]\n  \n  note: This error originates from a subprocess, and is likely not a problem with pip.\nerror: metadata-generation-failed\n\n\u00d7 Encountered error while generating package metadata.\n\u2570\u2500> See above for output.\n\nnote: This is an issue with the package mentioned above, not pip.\nhint: See above for details.\n
" ], "warnings": [ "W003: Missing classifier 'Framework :: AiiDA'", @@ -4273,7 +4273,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-ce
Collecting git+https://github.com/unkcpz/aiida-ce\n  Cloning https://github.com/unkcpz/aiida-ce to /tmp/pip-req-build-mpw7f9cf\n  Running command git clone --filter=blob:none --quiet https://github.com/unkcpz/aiida-ce /tmp/pip-req-build-mpw7f9cf\n  Resolved https://github.com/unkcpz/aiida-ce to commit 84cbc7c1e14af3cac5530e092e529bc84cfcfe6d\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Installing backend dependencies: started\n  Installing backend dependencies: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-ce to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-ce==0.1.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-ce 0.1.0 depends on aiida-core<2.0.0 and >=1.5.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-ce
Collecting git+https://github.com/unkcpz/aiida-ce\n  Cloning https://github.com/unkcpz/aiida-ce to /tmp/pip-req-build-dr7m931n\n  Running command git clone --filter=blob:none --quiet https://github.com/unkcpz/aiida-ce /tmp/pip-req-build-dr7m931n\n  Resolved https://github.com/unkcpz/aiida-ce to commit 84cbc7c1e14af3cac5530e092e529bc84cfcfe6d\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Installing backend dependencies: started\n  Installing backend dependencies: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-ce to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-ce==0.1.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-ce 0.1.0 depends on aiida-core<2.0.0 and >=1.5.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead." @@ -6552,7 +6552,7 @@ "verdi": "aiida.cmdline.commands.cmd_verdi:verdi" } }, - "commits_count": 280, + "commits_count": 277, "errors": [], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead." @@ -7864,7 +7864,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-crystal17
Collecting aiida-crystal17\n  Downloading aiida_crystal17-0.11.0-py3-none-any.whl.metadata (5.8 kB)\nINFO: pip is looking at multiple versions of aiida-crystal17 to determine which version is compatible with other requirements. This could take a while.\n  Downloading aiida-crystal17-0.4.0.tar.gz (120 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 120.8/120.8 kB 4.4 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nERROR: Cannot install aiida-crystal17==0.11.0 and aiida-crystal17==0.4.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-crystal17 0.11.0 depends on aiida-core<2.0.0 and >=1.4.0\n    aiida-crystal17 0.4.0 depends on aiida-core==0.12.2\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-crystal17
Collecting aiida-crystal17\n  Downloading aiida_crystal17-0.11.0-py3-none-any.whl.metadata (5.8 kB)\nINFO: pip is looking at multiple versions of aiida-crystal17 to determine which version is compatible with other requirements. This could take a while.\n  Downloading aiida-crystal17-0.4.0.tar.gz (120 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 120.8/120.8 kB 5.7 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nERROR: Cannot install aiida-crystal17==0.11.0 and aiida-crystal17==0.4.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-crystal17 0.11.0 depends on aiida-core<2.0.0 and >=1.4.0\n    aiida-crystal17 0.4.0 depends on aiida-core==0.12.2\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead." @@ -7957,7 +7957,7 @@ "commits_count": 0, "development_status": "beta", "errors": [ - "E001: Failed to install plugin aiida-cusp
Collecting https://pypi.org/project/aiida-cusp\n  Downloading https://pypi.org/project/aiida-cusp (43 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 43.3/43.3 kB 6.0 MB/s eta 0:00:00\n  ERROR: Cannot unpack file /tmp/pip-unpack-82c84079/aiida-cusp (downloaded from /tmp/pip-req-build-qws7j_vk, content-type: text/html; charset=UTF-8); cannot detect archive format\nERROR: Cannot determine archive format of /tmp/pip-req-build-qws7j_vk\n
" + "E001: Failed to install plugin aiida-cusp
Collecting https://pypi.org/project/aiida-cusp\n  Downloading https://pypi.org/project/aiida-cusp (43 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 43.3/43.3 kB 6.0 MB/s eta 0:00:00\n  ERROR: Cannot unpack file /tmp/pip-unpack-0scpd51m/aiida-cusp (downloaded from /tmp/pip-req-build-vrjss5ft, content-type: text/html; charset=UTF-8); cannot detect archive format\nERROR: Cannot determine archive format of /tmp/pip-req-build-vrjss5ft\n
" ], "warnings": [ "W010: Entry point 'potcar' does not start with prefix 'cusp.'" @@ -8329,7 +8329,7 @@ "commits_count": 0, "development_status": "beta", "errors": [ - "E003: Failed to fetch entry point metadata for package aiida_defects
Traceback (most recent call last):\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 196, in \n    cli()  # pylint: disable=no-value-for-parameter\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 1157, in __call__\nFailed to load entry point 'defects.formation_energy.corrections.point_countercharge':\nTraceback (most recent call last):\n  File \"/opt/conda/lib/python3.10/site-packages/aiida/plugins/entry_point.py\", line 278, in load_entry_point\n    loaded_entry_point = entry_point.load()\n  File \"/opt/conda/lib/python3.10/site-packages/importlib_metadata/__init__.py\", line 211, in load\n    module = import_module(match.group('module'))\n  File \"/opt/conda/lib/python3.10/importlib/__init__.py\", line 126, in import_module\n    return _bootstrap._gcd_import(name[level:], package, level)\n  File \"\", line 1050, in _gcd_import\n  File \"\", line 1027, in _find_and_load\n  File \"\", line 1006, in _find_and_load_unlocked\n  File \"\", line 688, in _load_unlocked\n  File \"\", line 883, in exec_module\n  File \"\", line 241, in _call_with_frames_removed\n  File \"/opt/conda/lib/python3.10/site-packages/aiida_defects/formation_energy/corrections/point_countercharge/point_countercharge.py\", line 19, in \n    from aiida_defects.tools.structure_manipulation import get_spacegroup\nModuleNotFoundError: No module named 'aiida_defects.tools'\n\n    return self.main(*args, **kwargs)\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 1078, in main\n    rv = self.invoke(ctx)\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 1434, in invoke\n    return ctx.invoke(self.callback, **ctx.params)\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 783, in invoke\n    return __callback(*args, **kwargs)\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 186, in cli\n    result[ep_group] = document_entry_point_group(ep_group)\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 75, in document_entry_point_group\n    process_info = document_entry_point(entry_point_group, entry_point)\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 94, in document_entry_point\n    plugin = load_entry_point(entry_point_group, entry_point)\n  File \"/opt/conda/lib/python3.10/site-packages/aiida/plugins/entry_point.py\", line 278, in load_entry_point\n    loaded_entry_point = entry_point.load()\n  File \"/opt/conda/lib/python3.10/site-packages/importlib_metadata/__init__.py\", line 213, in load\n    return functools.reduce(getattr, attrs, module)\nAttributeError: module 'aiida_defects.formation_energy.formation_energy_siesta' has no attribute 'FormatonEnergyWorkchainSiesta'. Did you mean: 'FormationEnergyWorkchainSiesta'?\n
" + "E003: Failed to fetch entry point metadata for package aiida_defects
Traceback (most recent call last):\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 196, in \n    cli()  # pylint: disable=no-value-for-parameter\nFailed to load entry point 'defects.formation_energy.corrections.point_countercharge':\nTraceback (most recent call last):\n  File \"/opt/conda/lib/python3.10/site-packages/aiida/plugins/entry_point.py\", line 278, in load_entry_point\n    loaded_entry_point = entry_point.load()\n  File \"/opt/conda/lib/python3.10/site-packages/importlib_metadata/__init__.py\", line 211, in load\n    module = import_module(match.group('module'))\n  File \"/opt/conda/lib/python3.10/importlib/__init__.py\", line 126, in import_module\n    return _bootstrap._gcd_import(name[level:], package, level)\n  File \"\", line 1050, in _gcd_import\n  File \"\", line 1027, in _find_and_load\n  File \"\", line 1006, in _find_and_load_unlocked\n  File \"\", line 688, in _load_unlocked\n  File \"\", line 883, in exec_module\n  File \"\", line 241, in _call_with_frames_removed\n  File \"/opt/conda/lib/python3.10/site-packages/aiida_defects/formation_energy/corrections/point_countercharge/point_countercharge.py\", line 19, in \n    from aiida_defects.tools.structure_manipulation import get_spacegroup\nModuleNotFoundError: No module named 'aiida_defects.tools'\n\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 1157, in __call__\n    return self.main(*args, **kwargs)\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 1078, in main\n    rv = self.invoke(ctx)\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 1434, in invoke\n    return ctx.invoke(self.callback, **ctx.params)\n  File \"/opt/conda/lib/python3.10/site-packages/click/core.py\", line 783, in invoke\n    return __callback(*args, **kwargs)\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 186, in cli\n    result[ep_group] = document_entry_point_group(ep_group)\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 75, in document_entry_point_group\n    process_info = document_entry_point(entry_point_group, entry_point)\n  File \"/tmp/scripts/./bin/analyze_entrypoints.py\", line 94, in document_entry_point\n    plugin = load_entry_point(entry_point_group, entry_point)\n  File \"/opt/conda/lib/python3.10/site-packages/aiida/plugins/entry_point.py\", line 278, in load_entry_point\n    loaded_entry_point = entry_point.load()\n  File \"/opt/conda/lib/python3.10/site-packages/importlib_metadata/__init__.py\", line 213, in load\n    return functools.reduce(getattr, attrs, module)\nAttributeError: module 'aiida_defects.formation_energy.formation_energy_siesta' has no attribute 'FormatonEnergyWorkchainSiesta'. Did you mean: 'FormationEnergyWorkchainSiesta'?\n
" ], "warnings": [], "summaryinfo": [ @@ -8385,7 +8385,7 @@ }, "commits_count": 2, "errors": [ - "E001: Failed to install plugin aiida-diff
Collecting aiida-diff-0.1.0a0\n  Cloning https://github.com/aiidateam/aiida-diff to /tmp/pip-install-t2bsmeet/aiida-diff-0-1-0a0_882844d476434bebb0e95d1859fb1bf3\n  Running command git clone --filter=blob:none --quiet https://github.com/aiidateam/aiida-diff /tmp/pip-install-t2bsmeet/aiida-diff-0-1-0a0_882844d476434bebb0e95d1859fb1bf3\n  Resolved https://github.com/aiidateam/aiida-diff to commit f9b151e11634b16e6a42d5445da23735415872e0\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\n  WARNING: Generating metadata for package aiida-diff-0.1.0a0 produced metadata for project name aiida-diff. Fix your #egg=aiida-diff-0.1.0a0 fragments.\nDiscarding git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0: Requested aiida-diff from git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0 has inconsistent name: expected 'aiida-diff-0-1-0a0', but metadata has 'aiida-diff'\nERROR: Could not find a version that satisfies the requirement aiida-diff-0-1-0a0 (unavailable) (from versions: none)\nERROR: No matching distribution found for aiida-diff-0-1-0a0 (unavailable)\n
" + "E001: Failed to install plugin aiida-diff
Collecting aiida-diff-0.1.0a0\n  Cloning https://github.com/aiidateam/aiida-diff to /tmp/pip-install-ps0wxtvo/aiida-diff-0-1-0a0_cdad228f0c3e482f899eab94d021c5d2\n  Running command git clone --filter=blob:none --quiet https://github.com/aiidateam/aiida-diff /tmp/pip-install-ps0wxtvo/aiida-diff-0-1-0a0_cdad228f0c3e482f899eab94d021c5d2\n  Resolved https://github.com/aiidateam/aiida-diff to commit f9b151e11634b16e6a42d5445da23735415872e0\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\n  WARNING: Generating metadata for package aiida-diff-0.1.0a0 produced metadata for project name aiida-diff. Fix your #egg=aiida-diff-0.1.0a0 fragments.\nDiscarding git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0: Requested aiida-diff from git+https://github.com/aiidateam/aiida-diff#egg=aiida-diff-0.1.0a0 has inconsistent name: expected 'aiida-diff-0-1-0a0', but metadata has 'aiida-diff'\nERROR: Could not find a version that satisfies the requirement aiida-diff-0-1-0a0 (unavailable) (from versions: none)\nERROR: No matching distribution found for aiida-diff-0-1-0a0 (unavailable)\n
" ], "warnings": [ "W005: Development status in classifiers (alpha) does not match development_status in metadata (stable)", @@ -9026,7 +9026,7 @@ "commits_count": 0, "development_status": "pre-alpha", "errors": [ - "E001: Failed to install plugin aiida-fenics
Collecting git+https://github.com/broeder-j/aiida-fenics\n  Cloning https://github.com/broeder-j/aiida-fenics to /tmp/pip-req-build-blbbjtcm\n  Running command git clone --filter=blob:none --quiet https://github.com/broeder-j/aiida-fenics /tmp/pip-req-build-blbbjtcm\n  Resolved https://github.com/broeder-j/aiida-fenics to commit eae053f7c82889a15fe19c3909d303f923fcf79e\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-fenics to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-fenics==0.1.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-fenics 0.1.0 depends on aiida-core~=1.6\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-fenics
Collecting git+https://github.com/broeder-j/aiida-fenics\n  Cloning https://github.com/broeder-j/aiida-fenics to /tmp/pip-req-build-u53c3kzm\n  Running command git clone --filter=blob:none --quiet https://github.com/broeder-j/aiida-fenics /tmp/pip-req-build-u53c3kzm\n  Resolved https://github.com/broeder-j/aiida-fenics to commit eae053f7c82889a15fe19c3909d303f923fcf79e\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-fenics to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-fenics==0.1.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-fenics 0.1.0 depends on aiida-core~=1.6\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [], "summaryinfo": [ @@ -13552,7 +13552,7 @@ "commits_count": 0, "development_status": "alpha", "errors": [ - "E001: Failed to install plugin aiida-graphql
Collecting aiida-graphql\n  Downloading aiida_graphql-0.0.2-py3-none-any.whl.metadata (3.3 kB)\nCollecting aiida<2.0.0,>=1.0.0b6 (from aiida-graphql)\n  Downloading aiida-1.0.1.tar.gz (2.8 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.10-py3-none-any.whl.metadata (4.0 kB)\nRequirement already satisfied: aiida-core in /opt/conda/lib/python3.10/site-packages (from aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.5.2)\nCollecting click<8.0,>=7.0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading click-7.1.2-py2.py3-none-any.whl.metadata (2.9 kB)\nCollecting graphql-core<4.0.0,>=3.0.0a0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading graphql_core-3.3.0a6-py3-none-any.whl.metadata (11 kB)\nCollecting hupper<2.0,>=1.5 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading hupper-1.12.1-py3-none-any.whl.metadata (3.7 kB)\nRequirement already satisfied: pygments<3.0,>=2.3 in /opt/conda/lib/python3.10/site-packages (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql) (2.18.0)\nCollecting starlette==0.12.10 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading starlette-0.12.10.tar.gz (46 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 46.3/46.3 kB 4.9 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting uvicorn==0.10.0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvicorn-0.10.0.tar.gz (26 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting h11==0.8.* (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading h11-0.8.1-py2.py3-none-any.whl.metadata (8.0 kB)\nCollecting websockets==8.* (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading websockets-8.1.tar.gz (58 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 58.9/58.9 kB 14.9 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting httptools==0.0.13 (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading httptools-0.0.13.tar.gz (104 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 104.2/104.2 kB 24.6 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting uvloop==0.14.0rc2 (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvloop-0.14.0rc2.tar.gz (2.0 MB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.0/2.0 MB 70.1 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nRequirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.13.1)\nRequirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.4.2)\nRequirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.8.1)\nRequirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.18.0)\nRequirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.1.10)\nINFO: pip is looking at multiple versions of aiida-core to determine which version is compatible with other requirements. This could take a while.\nCollecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.9-py3-none-any.whl.metadata (4.0 kB)\nCollecting uvicorn==0.9.1 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvicorn-0.9.1.tar.gz (25 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting uvloop==0.* (from uvicorn==0.9.1->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvloop-0.21.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB)\nCollecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.8-py3-none-any.whl.metadata (4.0 kB)\n  Downloading strawberry_graphql-0.16.7-py3-none-any.whl.metadata (4.0 kB)\nCollecting aiida<2.0.0,>=1.0.0b6 (from aiida-graphql)\n  Downloading aiida-1.0.0.tar.gz (1.9 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida to determine which version is compatible with other requirements. This could take a while.\n  Downloading aiida-1.0.0b6.tar.gz (2.6 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting aiida-graphql\n  Downloading aiida_graphql-0.0.1-py3-none-any.whl.metadata (3.4 kB)\nRequirement already satisfied: click~=8.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (8.1.7)\nRequirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.1.0)\nRequirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.16)\nRequirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.20.3)\nRequirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (8.25.0)\nRequirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.18.2)\nRequirement already satisfied: jinja2~=3.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (3.1.4)\nRequirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.7.8)\nRequirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.11.0)\nRequirement already satisfied: numpy~=1.21 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.26.4)\nRequirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.12.0)\nRequirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.21.10)\nRequirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.2.4)\nRequirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (5.9.8)\nRequirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.9.9)\nRequirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.7.3)\nRequirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2021.3)\nRequirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.0.1)\nRequirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.32.3)\nRequirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.0.30)\nRequirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.8.10)\nRequirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (4.66.4)\nRequirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.9.5)\nRequirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.16.0)\nINFO: pip is looking at multiple versions of strawberry-graphql to determine which version is compatible with other requirements. This could take a while.\nCollecting strawberry-graphql<0.17.0,>=0.16.5 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.6-py3-none-any.whl.metadata (4.0 kB)\n  Downloading strawberry_graphql-0.16.5-py3-none-any.whl.metadata (4.0 kB)\nCollecting aiida-core (from aiida<2.0.0,>=1.0.0b6->aiida-graphql)\n  Downloading aiida_core-2.5.2-py3-none-any.whl.metadata (11 kB)\nINFO: pip is still looking at multiple versions of aiida-core to determine which version is compatible with other requirements. This could take a while.\nCollecting uvicorn==0.9.0 (from strawberry-graphql<0.17.0,>=0.16.5->aiida-graphql)\n  Downloading uvicorn-0.9.0.tar.gz (24 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nERROR: Cannot install aiida-graphql because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida 1.0.0 depends on aiida-core==1.0\n    aiida 1.0.0b6 depends on aiida-core==1.0.0b6\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-graphql
Collecting aiida-graphql\n  Downloading aiida_graphql-0.0.2-py3-none-any.whl.metadata (3.3 kB)\nCollecting aiida<2.0.0,>=1.0.0b6 (from aiida-graphql)\n  Downloading aiida-1.0.1.tar.gz (2.8 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.10-py3-none-any.whl.metadata (4.0 kB)\nRequirement already satisfied: aiida-core in /opt/conda/lib/python3.10/site-packages (from aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.5.2)\nCollecting click<8.0,>=7.0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading click-7.1.2-py2.py3-none-any.whl.metadata (2.9 kB)\nCollecting graphql-core<4.0.0,>=3.0.0a0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading graphql_core-3.3.0a6-py3-none-any.whl.metadata (11 kB)\nCollecting hupper<2.0,>=1.5 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading hupper-1.12.1-py3-none-any.whl.metadata (3.7 kB)\nRequirement already satisfied: pygments<3.0,>=2.3 in /opt/conda/lib/python3.10/site-packages (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql) (2.18.0)\nCollecting starlette==0.12.10 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading starlette-0.12.10.tar.gz (46 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 46.3/46.3 kB 3.8 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting uvicorn==0.10.0 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvicorn-0.10.0.tar.gz (26 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting h11==0.8.* (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading h11-0.8.1-py2.py3-none-any.whl.metadata (8.0 kB)\nCollecting websockets==8.* (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading websockets-8.1.tar.gz (58 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 58.9/58.9 kB 14.8 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting httptools==0.0.13 (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading httptools-0.0.13.tar.gz (104 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 104.2/104.2 kB 19.7 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting uvloop==0.14.0rc2 (from uvicorn==0.10.0->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvloop-0.14.0rc2.tar.gz (2.0 MB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.0/2.0 MB 12.1 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nRequirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.13.1)\nRequirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.4.2)\nRequirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.8.1)\nRequirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.18.0)\nRequirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.1.10)\nINFO: pip is looking at multiple versions of aiida-core to determine which version is compatible with other requirements. This could take a while.\nCollecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.9-py3-none-any.whl.metadata (4.0 kB)\nCollecting uvicorn==0.9.1 (from strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvicorn-0.9.1.tar.gz (25 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting uvloop==0.* (from uvicorn==0.9.1->strawberry-graphql<0.17.0,>=0.16.7->aiida-graphql)\n  Downloading uvloop-0.21.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (4.9 kB)\nCollecting strawberry-graphql<0.17.0,>=0.16.7 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.8-py3-none-any.whl.metadata (4.0 kB)\n  Downloading strawberry_graphql-0.16.7-py3-none-any.whl.metadata (4.0 kB)\nCollecting aiida<2.0.0,>=1.0.0b6 (from aiida-graphql)\n  Downloading aiida-1.0.0.tar.gz (1.9 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida to determine which version is compatible with other requirements. This could take a while.\n  Downloading aiida-1.0.0b6.tar.gz (2.6 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting aiida-graphql\n  Downloading aiida_graphql-0.0.1-py3-none-any.whl.metadata (3.4 kB)\nRequirement already satisfied: click~=8.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (8.1.7)\nRequirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.1.0)\nRequirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.16)\nRequirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.20.3)\nRequirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (8.25.0)\nRequirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.18.2)\nRequirement already satisfied: jinja2~=3.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (3.1.4)\nRequirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.7.8)\nRequirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.11.0)\nRequirement already satisfied: numpy~=1.21 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.26.4)\nRequirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.12.0)\nRequirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.21.10)\nRequirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.2.4)\nRequirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (5.9.8)\nRequirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.9.9)\nRequirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.7.3)\nRequirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2021.3)\nRequirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (6.0.1)\nRequirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.32.3)\nRequirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (2.0.30)\nRequirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.8.10)\nRequirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (4.66.4)\nRequirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (0.9.5)\nRequirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core->aiida<2.0.0,>=1.0.0b6->aiida-graphql) (1.16.0)\nINFO: pip is looking at multiple versions of strawberry-graphql to determine which version is compatible with other requirements. This could take a while.\nCollecting strawberry-graphql<0.17.0,>=0.16.5 (from aiida-graphql)\n  Downloading strawberry_graphql-0.16.6-py3-none-any.whl.metadata (4.0 kB)\n  Downloading strawberry_graphql-0.16.5-py3-none-any.whl.metadata (4.0 kB)\nCollecting aiida-core (from aiida<2.0.0,>=1.0.0b6->aiida-graphql)\n  Downloading aiida_core-2.5.2-py3-none-any.whl.metadata (11 kB)\nINFO: pip is still looking at multiple versions of aiida-core to determine which version is compatible with other requirements. This could take a while.\nCollecting uvicorn==0.9.0 (from strawberry-graphql<0.17.0,>=0.16.5->aiida-graphql)\n  Downloading uvicorn-0.9.0.tar.gz (24 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nERROR: Cannot install aiida-graphql because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida 1.0.0 depends on aiida-core==1.0\n    aiida 1.0.0b6 depends on aiida-core==1.0.0b6\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W020: Unable to read wheel file from PyPI release:
No entry_points.txt found in wheel
", @@ -20421,7 +20421,7 @@ } } }, - "commits_count": 61, + "commits_count": 60, "errors": [], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead." @@ -20497,7 +20497,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-lsmo
Collecting git+https://github.com/lsmo-epfl/aiida-lsmo\n  Cloning https://github.com/lsmo-epfl/aiida-lsmo to /tmp/pip-req-build-jclaqs35\n  Running command git clone --filter=blob:none --quiet https://github.com/lsmo-epfl/aiida-lsmo /tmp/pip-req-build-jclaqs35\n  Resolved https://github.com/lsmo-epfl/aiida-lsmo to commit 6bf08fa42e545dadf889ea8095d7fcdd8d1be15c\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-lsmo to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-lsmo==1.0.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-lsmo 1.0.0 depends on aiida-core~=1.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-lsmo
Collecting git+https://github.com/lsmo-epfl/aiida-lsmo\n  Cloning https://github.com/lsmo-epfl/aiida-lsmo to /tmp/pip-req-build-m59d81jw\n  Running command git clone --filter=blob:none --quiet https://github.com/lsmo-epfl/aiida-lsmo /tmp/pip-req-build-m59d81jw\n  Resolved https://github.com/lsmo-epfl/aiida-lsmo to commit 6bf08fa42e545dadf889ea8095d7fcdd8d1be15c\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-lsmo to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-lsmo==1.0.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-lsmo 1.0.0 depends on aiida-core~=1.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W003: Missing classifier 'Framework :: AiiDA'", @@ -20561,7 +20561,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-metavo-scheduler
Collecting git+https://github.com/pzarabadip/aiida-metavo-scheduler\n  Cloning https://github.com/pzarabadip/aiida-metavo-scheduler to /tmp/pip-req-build-1hk9jpjb\n  Running command git clone --filter=blob:none --quiet https://github.com/pzarabadip/aiida-metavo-scheduler /tmp/pip-req-build-1hk9jpjb\n  Resolved https://github.com/pzarabadip/aiida-metavo-scheduler to commit 955697497641ca13e997431a8e925df6ec3a9eea\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-metavo-scheduler to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-metavo-scheduler==1.0.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-metavo-scheduler 1.0.0 depends on aiida_core<1.6 and >=1.0.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-metavo-scheduler
Collecting git+https://github.com/pzarabadip/aiida-metavo-scheduler\n  Cloning https://github.com/pzarabadip/aiida-metavo-scheduler to /tmp/pip-req-build-g6epo121\n  Running command git clone --filter=blob:none --quiet https://github.com/pzarabadip/aiida-metavo-scheduler /tmp/pip-req-build-g6epo121\n  Resolved https://github.com/pzarabadip/aiida-metavo-scheduler to commit 955697497641ca13e997431a8e925df6ec3a9eea\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-metavo-scheduler to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-metavo-scheduler==1.0.0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-metavo-scheduler 1.0.0 depends on aiida_core<1.6 and >=1.0.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.", @@ -21488,7 +21488,7 @@ }, "commits_count": 4, "errors": [ - "E001: Failed to install plugin aiida-mpds
Collecting git+https://github.com/mpds-io/mpds-aiida\n  Cloning https://github.com/mpds-io/mpds-aiida to /tmp/pip-req-build-rly92ly4\n  Running command git clone --filter=blob:none --quiet https://github.com/mpds-io/mpds-aiida /tmp/pip-req-build-rly92ly4\n  Resolved https://github.com/mpds-io/mpds-aiida to commit b4ab96db1472969e89159b2b2d76a9836b99c94a\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs (from mpds-aiida==0.10.1)\n  Cloning https://github.com/mpds-io/mpds-ml-labs to /tmp/pip-install-3cmc92_1/mpds-ml-labs_d03ad950d69e47fea3a08632b4278c5d\n  Running command git clone --filter=blob:none --quiet https://github.com/mpds-io/mpds-ml-labs /tmp/pip-install-3cmc92_1/mpds-ml-labs_d03ad950d69e47fea3a08632b4278c5d\n  Resolved https://github.com/mpds-io/mpds-ml-labs to commit 46b15d3fb27a1ca3389d9fbe44b9061740291bc7\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft (from mpds-aiida==0.10.1)\n  Cloning https://github.com/tilde-lab/aiida-crystal-dft to /tmp/pip-install-3cmc92_1/aiida-crystal-dft_9f967b87afb146b898d725b6108bb3a4\n  Running command git clone --filter=blob:none --quiet https://github.com/tilde-lab/aiida-crystal-dft /tmp/pip-install-3cmc92_1/aiida-crystal-dft_9f967b87afb146b898d725b6108bb3a4\n  Resolved https://github.com/tilde-lab/aiida-crystal-dft to commit dfa54170debe4be80c9fc8983e826af46afe5916\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nCollecting reentry>=1.3.1 (from mpds-aiida==0.10.1)\n  Using cached reentry-1.3.3-py3-none-any.whl.metadata (10 kB)\nRequirement already satisfied: six>=1.16.0 in /opt/conda/lib/python3.10/site-packages (from mpds-aiida==0.10.1) (1.16.0)\nRequirement already satisfied: numpy>=1.17.5 in /opt/conda/lib/python3.10/site-packages (from mpds-aiida==0.10.1) (1.26.4)\nCollecting ase>=3.19 (from mpds-aiida==0.10.1)\n  Downloading ase-3.23.0-py3-none-any.whl.metadata (3.8 kB)\nCollecting yascheduler>=1.0.12 (from mpds-aiida==0.10.1)\n  Downloading yascheduler-1.2.0-py3-none-any.whl.metadata (12 kB)\nCollecting scipy>=1.6.0 (from ase>=3.19->mpds-aiida==0.10.1)\n  Downloading scipy-1.14.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (60 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 60.8/60.8 kB 5.1 MB/s eta 0:00:00\nCollecting matplotlib>=3.3.4 (from ase>=3.19->mpds-aiida==0.10.1)\n  Downloading matplotlib-3.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB)\nRequirement already satisfied: setuptools>=36.2 in /opt/conda/lib/python3.10/site-packages (from reentry>=1.3.1->mpds-aiida==0.10.1) (70.0.0)\nRequirement already satisfied: click in /opt/conda/lib/python3.10/site-packages (from reentry>=1.3.1->mpds-aiida==0.10.1) (8.1.7)\nCollecting aiohttp~=3.8 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading aiohttp-3.11.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)\nCollecting asyncssh~=2.11 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading asyncssh-2.18.0-py3-none-any.whl.metadata (9.9 kB)\nCollecting asyncstdlib~=3.10 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading asyncstdlib-3.13.0-py3-none-any.whl.metadata (3.7 kB)\nCollecting attrs~=21.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading attrs-21.4.0-py2.py3-none-any.whl.metadata (9.8 kB)\nCollecting azure-identity~=1.10.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_identity-1.10.0-py3-none-any.whl.metadata (54 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 54.5/54.5 kB 15.7 MB/s eta 0:00:00\nCollecting azure-mgmt-compute~=27.2.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_mgmt_compute-27.2.0-py3-none-any.whl.metadata (53 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.1/53.1 kB 12.8 MB/s eta 0:00:00\nCollecting azure-mgmt-network~=20.0.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_mgmt_network-20.0.0-py3-none-any.whl.metadata (63 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 63.9/63.9 kB 21.5 MB/s eta 0:00:00\nCollecting backoff~=2.1.2 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading backoff-2.1.2-py3-none-any.whl.metadata (13 kB)\nCollecting hcloud~=1.17 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading hcloud-1.35.0-py3-none-any.whl.metadata (4.9 kB)\nCollecting pg8000~=1.19 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading pg8000-1.31.2-py3-none-any.whl.metadata (74 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 74.7/74.7 kB 16.7 MB/s eta 0:00:00\nCollecting python-daemon~=2.3 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading python_daemon-2.3.2-py3-none-any.whl.metadata (2.0 kB)\nRequirement already satisfied: typing-extensions>=4.2.0 in /opt/conda/lib/python3.10/site-packages (from yascheduler>=1.0.12->mpds-aiida==0.10.1) (4.12.1)\nCollecting upcloud_api~=2.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading upcloud_api-2.6.0-py3-none-any.whl.metadata (7.8 kB)\nRequirement already satisfied: aiida-core>=2.0.2 in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.5.2)\nCollecting pycrystal>=1.0.10 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pycrystal-1.0.16.tar.gz (28 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting mpds_client>=0.24 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading mpds_client-0.24.tar.gz (9.7 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting pyparsing>2.3.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pyparsing-3.2.0-py3-none-any.whl.metadata (5.0 kB)\nCollecting spglib==2.0.2 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading spglib-2.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.3 kB)\nRequirement already satisfied: jinja2>=2.10 in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.1.4)\nCollecting jsonschema>=3.0.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading jsonschema-4.23.0-py3-none-any.whl.metadata (7.9 kB)\nRequirement already satisfied: packaging in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (24.0)\nCollecting pycodcif (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading pycodcif-3.0.1.tar.gz (79 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 79.6/79.6 kB 15.1 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting scikit-learn (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading scikit_learn-1.5.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (13 kB)\nCollecting imblearn (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading imblearn-0.0-py2.py3-none-any.whl.metadata (355 bytes)\nCollecting progressbar (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading progressbar-2.5.tar.gz (10 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nRequirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.13.1)\nRequirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.4.2)\nRequirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.8.1)\nRequirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.18.0)\nRequirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.1.10)\nRequirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.1.0)\nRequirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.16)\nRequirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.20.3)\nRequirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (8.25.0)\nRequirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.18.2)\nRequirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.8)\nRequirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.11.0)\nRequirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.12.0)\nRequirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.21.10)\nRequirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.4)\nRequirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.9.8)\nRequirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.9.9)\nRequirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.7.3)\nRequirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2021.3)\nRequirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.0.1)\nRequirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.32.3)\nRequirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.0.30)\nRequirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.8.10)\nRequirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.66.4)\nRequirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.9.5)\nRequirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.16.0)\nCollecting aiohappyeyeballs>=2.3.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading aiohappyeyeballs-2.4.3-py3-none-any.whl.metadata (6.1 kB)\nCollecting aiosignal>=1.1.2 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading aiosignal-1.3.1-py3-none-any.whl.metadata (4.0 kB)\nCollecting async-timeout<6.0,>=4.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading async_timeout-5.0.1-py3-none-any.whl.metadata (5.1 kB)\nCollecting frozenlist>=1.1.1 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (13 kB)\nRequirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/lib/python3.10/site-packages (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1) (6.0.5)\nCollecting propcache>=0.2.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading propcache-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)\nCollecting yarl<2.0,>=1.17.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading yarl-1.18.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (67 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 67.7/67.7 kB 19.1 MB/s eta 0:00:00\nRequirement already satisfied: cryptography>=39.0 in /opt/conda/lib/python3.10/site-packages (from asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (42.0.8)\nCollecting azure-core<2.0.0,>=1.11.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_core-1.32.0-py3-none-any.whl.metadata (39 kB)\nCollecting msal<2.0.0,>=1.12.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading msal-1.31.1-py3-none-any.whl.metadata (11 kB)\nCollecting msal-extensions<2.0.0,>=0.3.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading msal_extensions-1.2.0-py3-none-any.whl.metadata (7.6 kB)\nCollecting msrest>=0.6.21 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading msrest-0.7.1-py3-none-any.whl.metadata (21 kB)\nCollecting azure-common~=1.1 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_common-1.1.28-py2.py3-none-any.whl.metadata (5.0 kB)\nCollecting azure-mgmt-core<2.0.0,>=1.3.1 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_mgmt_core-1.5.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting python-dateutil>=2.7.5 (from hcloud~=1.17->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl.metadata (8.4 kB)\nRequirement already satisfied: MarkupSafe>=2.0 in /opt/conda/lib/python3.10/site-packages (from jinja2>=2.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.1.5)\nINFO: pip is looking at multiple versions of jsonschema to determine which version is compatible with other requirements. This could take a while.\nCollecting jsonschema>=3.0.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading jsonschema-4.22.0-py3-none-any.whl.metadata (8.2 kB)\n  Downloading jsonschema-4.21.1-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.21.0-py3-none-any.whl.metadata (8.0 kB)\n  Downloading jsonschema-4.20.0-py3-none-any.whl.metadata (8.1 kB)\n  Downloading jsonschema-4.19.2-py3-none-any.whl.metadata (7.9 kB)\n  Downloading jsonschema-4.19.1-py3-none-any.whl.metadata (7.9 kB)\n  Downloading jsonschema-4.19.0-py3-none-any.whl.metadata (8.2 kB)\nINFO: pip is still looking at multiple versions of jsonschema to determine which version is compatible with other requirements. This could take a while.\n  Downloading jsonschema-4.18.6-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.18.5-py3-none-any.whl.metadata (7.7 kB)\n  Downloading jsonschema-4.18.4-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.18.3-py3-none-any.whl.metadata (7.9 kB)\n  Downloading jsonschema-4.18.2-py3-none-any.whl.metadata (7.8 kB)\nINFO: This is taking longer than usual. You might need to provide the dependency resolver with stricter constraints to reduce runtime. See https://pip.pypa.io/warnings/backtracking for guidance. If you want to abort this run, press Ctrl + C.\n  Downloading jsonschema-4.18.1-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.18.0-py3-none-any.whl.metadata (10 kB)\n  Downloading jsonschema-4.17.3-py3-none-any.whl.metadata (7.9 kB)\nCollecting pyrsistent!=0.17.0,!=0.17.1,!=0.17.2,>=0.14.0 (from jsonschema>=3.0.1->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pyrsistent-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (27 kB)\nCollecting contourpy>=1.0.1 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading contourpy-1.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.4 kB)\nCollecting cycler>=0.10 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading cycler-0.12.1-py3-none-any.whl.metadata (3.8 kB)\nCollecting fonttools>=4.22.0 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading fonttools-4.55.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (164 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 164.5/164.5 kB 38.6 MB/s eta 0:00:00\nCollecting kiwisolver>=1.3.1 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading kiwisolver-1.4.7-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl.metadata (6.3 kB)\nCollecting pillow>=8 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl.metadata (9.1 kB)\nCollecting httplib2 (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading httplib2-0.22.0-py3-none-any.whl.metadata (2.6 kB)\nCollecting ujson (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading ujson-5.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (9.3 kB)\nCollecting pandas (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (89 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 89.9/89.9 kB 11.6 MB/s eta 0:00:00\nCollecting jmespath (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading jmespath-1.0.1-py3-none-any.whl.metadata (7.6 kB)\nCollecting scramp>=1.4.5 (from pg8000~=1.19->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading scramp-1.4.5-py3-none-any.whl.metadata (19 kB)\nCollecting bs4 (from pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading bs4-0.0.2-py2.py3-none-any.whl.metadata (411 bytes)\nCollecting docutils (from python-daemon~=2.3->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading docutils-0.21.2-py3-none-any.whl.metadata (2.8 kB)\nCollecting lockfile>=0.10 (from python-daemon~=2.3->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading lockfile-0.12.2-py2.py3-none-any.whl.metadata (2.4 kB)\nCollecting imbalanced-learn (from imblearn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading imbalanced_learn-0.12.4-py3-none-any.whl.metadata (8.3 kB)\nCollecting joblib>=1.2.0 (from scikit-learn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading joblib-1.4.2-py3-none-any.whl.metadata (5.4 kB)\nCollecting threadpoolctl>=3.1.0 (from scikit-learn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading threadpoolctl-3.5.0-py3-none-any.whl.metadata (13 kB)\nRequirement already satisfied: aiormq<4,>=3.2.3 in /opt/conda/lib/python3.10/site-packages (from aio-pika~=6.6->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.3.1)\nRequirement already satisfied: Mako in /opt/conda/lib/python3.10/site-packages (from alembic~=1.2->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.3.5)\nRequirement already satisfied: pyzmq>=17.0 in /opt/conda/lib/python3.10/site-packages (from circus~=0.18.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (26.0.3)\nRequirement already satisfied: tornado>=5.0.2 in /opt/conda/lib/python3.10/site-packages (from circus~=0.18.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.4)\nRequirement already satisfied: cffi>=1.12 in /opt/conda/lib/python3.10/site-packages (from cryptography>=39.0->asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (1.16.0)\nRequirement already satisfied: zipp>=0.5 in /opt/conda/lib/python3.10/site-packages (from importlib-metadata~=6.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.19.2)\nRequirement already satisfied: decorator in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.1.1)\nRequirement already satisfied: matplotlib-inline in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.1.7)\nRequirement already satisfied: prompt-toolkit<3.1.0,>=3.0.41 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.0.46)\nRequirement already satisfied: pygments>=2.4.0 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.18.0)\nRequirement already satisfied: stack-data in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.6.3)\nRequirement already satisfied: traitlets>=5.13.0 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.14.3)\nRequirement already satisfied: exceptiongroup in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.2.1)\nRequirement already satisfied: pexpect>4.3 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.9.0)\nRequirement already satisfied: parso<0.9.0,>=0.8.0 in /opt/conda/lib/python3.10/site-packages (from jedi<0.19->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.8.4)\nRequirement already satisfied: shortuuid in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.0.13)\nRequirement already satisfied: async-generator in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.10)\nRequirement already satisfied: pytray<0.4.0,>=0.2.2 in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.3.4)\nRequirement already satisfied: deprecation in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.1.0)\nRequirement already satisfied: pamqp~=2.0 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.3.0)\nCollecting PyJWT<3,>=1.0.0 (from PyJWT[crypto]<3,>=1.0.0->msal<2.0.0,>=1.12.0->azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading PyJWT-2.10.0-py3-none-any.whl.metadata (4.0 kB)\nCollecting portalocker<3,>=1.4 (from msal-extensions<2.0.0,>=0.3.0->azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading portalocker-2.10.1-py3-none-any.whl.metadata (8.5 kB)\nRequirement already satisfied: certifi>=2017.4.17 in /opt/conda/lib/python3.10/site-packages (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1) (2024.2.2)\nCollecting isodate>=0.6.0 (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading isodate-0.7.2-py3-none-any.whl.metadata (11 kB)\nCollecting requests-oauthlib>=0.5.0 (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading requests_oauthlib-2.0.0-py2.py3-none-any.whl.metadata (11 kB)\nRequirement already satisfied: bcrypt>=3.1.3 in /opt/conda/lib/python3.10/site-packages (from paramiko>=2.7.2,~=2.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.1.3)\nRequirement already satisfied: pynacl>=1.0.1 in /opt/conda/lib/python3.10/site-packages (from paramiko>=2.7.2,~=2.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.5.0)\nRequirement already satisfied: nest_asyncio~=1.5 in /opt/conda/lib/python3.10/site-packages (from plumpy~=0.21.6->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.6.0)\nRequirement already satisfied: annotated-types>=0.4.0 in /opt/conda/lib/python3.10/site-packages (from pydantic~=2.4->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.0)\nRequirement already satisfied: pydantic-core==2.18.4 in /opt/conda/lib/python3.10/site-packages (from pydantic~=2.4->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.18.4)\nRequirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.3.2)\nRequirement already satisfied: idna<4,>=2.5 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.7)\nRequirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.2.1)\nCollecting asn1crypto>=1.5.1 (from scramp>=1.4.5->pg8000~=1.19->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading asn1crypto-1.5.1-py2.py3-none-any.whl.metadata (13 kB)\nRequirement already satisfied: greenlet!=0.4.17 in /opt/conda/lib/python3.10/site-packages (from sqlalchemy~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.0.3)\nCollecting beautifulsoup4 (from bs4->pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading beautifulsoup4-4.12.3-py3-none-any.whl.metadata (3.8 kB)\nCollecting tzdata>=2022.7 (from pandas->mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading tzdata-2024.2-py2.py3-none-any.whl.metadata (1.4 kB)\nRequirement already satisfied: pycparser in /opt/conda/lib/python3.10/site-packages (from cffi>=1.12->cryptography>=39.0->asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (2.22)\nRequirement already satisfied: ptyprocess>=0.5 in /opt/conda/lib/python3.10/site-packages (from pexpect>4.3->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.0)\nRequirement already satisfied: wcwidth in /opt/conda/lib/python3.10/site-packages (from prompt-toolkit<3.1.0,>=3.0.41->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.13)\nCollecting oauthlib>=3.0.0 (from requests-oauthlib>=0.5.0->msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading oauthlib-3.2.2-py3-none-any.whl.metadata (7.5 kB)\nCollecting soupsieve>1.2 (from beautifulsoup4->bs4->pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading soupsieve-2.6-py3-none-any.whl.metadata (4.6 kB)\nRequirement already satisfied: executing>=1.2.0 in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.0.1)\nRequirement already satisfied: asttokens>=2.1.0 in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.4.1)\nRequirement already satisfied: pure-eval in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.2)\nDownloading ase-3.23.0-py3-none-any.whl (2.9 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.9/2.9 MB 94.0 MB/s eta 0:00:00\nUsing cached reentry-1.3.3-py3-none-any.whl (17 kB)\nDownloading yascheduler-1.2.0-py3-none-any.whl (67 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 67.9/67.9 kB 13.9 MB/s eta 0:00:00\nDownloading spglib-2.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (515 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 515.3/515.3 kB 75.5 MB/s eta 0:00:00\nDownloading aiohttp-3.11.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.6 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.6/1.6 MB 99.2 MB/s eta 0:00:00\nDownloading asyncssh-2.18.0-py3-none-any.whl (367 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 367.6/367.6 kB 52.3 MB/s eta 0:00:00\nDownloading asyncstdlib-3.13.0-py3-none-any.whl (43 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 43.9/43.9 kB 11.9 MB/s eta 0:00:00\nDownloading attrs-21.4.0-py2.py3-none-any.whl (60 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 60.6/60.6 kB 17.0 MB/s eta 0:00:00\nDownloading azure_identity-1.10.0-py3-none-any.whl (134 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 134.1/134.1 kB 34.7 MB/s eta 0:00:00\nDownloading azure_mgmt_compute-27.2.0-py3-none-any.whl (5.7 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 5.7/5.7 MB 109.7 MB/s eta 0:00:00\nDownloading azure_mgmt_network-20.0.0-py3-none-any.whl (8.5 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 8.5/8.5 MB 104.1 MB/s eta 0:00:00\nDownloading backoff-2.1.2-py3-none-any.whl (14 kB)\nDownloading hcloud-1.35.0-py3-none-any.whl (85 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 85.4/85.4 kB 14.5 MB/s eta 0:00:00\nDownloading jsonschema-4.17.3-py3-none-any.whl (90 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 90.4/90.4 kB 16.2 MB/s eta 0:00:00\nDownloading matplotlib-3.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.3 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 8.3/8.3 MB 93.2 MB/s eta 0:00:00\nDownloading pg8000-1.31.2-py3-none-any.whl (54 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 54.5/54.5 kB 12.6 MB/s eta 0:00:00\nDownloading pyparsing-3.2.0-py3-none-any.whl (106 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 106.9/106.9 kB 30.8 MB/s eta 0:00:00\nDownloading python_daemon-2.3.2-py3-none-any.whl (34 kB)\nDownloading scipy-1.14.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (41.2 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 41.2/41.2 MB 68.8 MB/s eta 0:00:00\nDownloading upcloud_api-2.6.0-py3-none-any.whl (37 kB)\nDownloading imblearn-0.0-py2.py3-none-any.whl (1.9 kB)\nDownloading scikit_learn-1.5.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.3 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 13.3/13.3 MB 112.2 MB/s eta 0:00:00\nDownloading aiohappyeyeballs-2.4.3-py3-none-any.whl (14 kB)\nDownloading aiosignal-1.3.1-py3-none-any.whl (7.6 kB)\nDownloading async_timeout-5.0.1-py3-none-any.whl (6.2 kB)\nDownloading azure_common-1.1.28-py2.py3-none-any.whl (14 kB)\nDownloading azure_core-1.32.0-py3-none-any.whl (198 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 198.9/198.9 kB 48.2 MB/s eta 0:00:00\nDownloading azure_mgmt_core-1.5.0-py3-none-any.whl (30 kB)\nDownloading contourpy-1.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (324 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 325.0/325.0 kB 70.7 MB/s eta 0:00:00\nDownloading cycler-0.12.1-py3-none-any.whl (8.3 kB)\nDownloading fonttools-4.55.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (4.6 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 4.6/4.6 MB 122.2 MB/s eta 0:00:00\nDownloading frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (241 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 241.9/241.9 kB 38.1 MB/s eta 0:00:00\nDownloading joblib-1.4.2-py3-none-any.whl (301 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 301.8/301.8 kB 47.4 MB/s eta 0:00:00\nDownloading kiwisolver-1.4.7-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (1.6 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.6/1.6 MB 81.2 MB/s eta 0:00:00\nDownloading lockfile-0.12.2-py2.py3-none-any.whl (13 kB)\nDownloading msal-1.31.1-py3-none-any.whl (113 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 113.2/113.2 kB 27.0 MB/s eta 0:00:00\nDownloading msal_extensions-1.2.0-py3-none-any.whl (19 kB)\nDownloading msrest-0.7.1-py3-none-any.whl (85 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 85.4/85.4 kB 21.3 MB/s eta 0:00:00\nDownloading pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl (4.4 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 4.4/4.4 MB 109.4 MB/s eta 0:00:00\nDownloading propcache-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (208 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 208.9/208.9 kB 43.2 MB/s eta 0:00:00\nDownloading pyrsistent-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (117 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 117.7/117.7 kB 32.5 MB/s eta 0:00:00\nDownloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 229.9/229.9 kB 53.5 MB/s eta 0:00:00\nDownloading scramp-1.4.5-py3-none-any.whl (12 kB)\nDownloading threadpoolctl-3.5.0-py3-none-any.whl (18 kB)\nDownloading yarl-1.18.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (319 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 319.4/319.4 kB 71.9 MB/s eta 0:00:00\nDownloading bs4-0.0.2-py2.py3-none-any.whl (1.2 kB)\nDownloading docutils-0.21.2-py3-none-any.whl (587 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 587.4/587.4 kB 87.4 MB/s eta 0:00:00\nDownloading httplib2-0.22.0-py3-none-any.whl (96 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 96.9/96.9 kB 34.4 MB/s eta 0:00:00\nDownloading imbalanced_learn-0.12.4-py3-none-any.whl (258 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 258.3/258.3 kB 54.6 MB/s eta 0:00:00\nDownloading jmespath-1.0.1-py3-none-any.whl (20 kB)\nDownloading pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.1 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 13.1/13.1 MB 107.1 MB/s eta 0:00:00\nDownloading ujson-5.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (53 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.6/53.6 kB 17.0 MB/s eta 0:00:00\nDownloading asn1crypto-1.5.1-py2.py3-none-any.whl (105 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 105.0/105.0 kB 34.1 MB/s eta 0:00:00\nDownloading isodate-0.7.2-py3-none-any.whl (22 kB)\nDownloading portalocker-2.10.1-py3-none-any.whl (18 kB)\nDownloading PyJWT-2.10.0-py3-none-any.whl (23 kB)\nDownloading requests_oauthlib-2.0.0-py2.py3-none-any.whl (24 kB)\nDownloading tzdata-2024.2-py2.py3-none-any.whl (346 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 346.6/346.6 kB 54.9 MB/s eta 0:00:00\nDownloading beautifulsoup4-4.12.3-py3-none-any.whl (147 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 147.9/147.9 kB 36.9 MB/s eta 0:00:00\nDownloading oauthlib-3.2.2-py3-none-any.whl (151 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 151.7/151.7 kB 30.4 MB/s eta 0:00:00\nDownloading soupsieve-2.6-py3-none-any.whl (36 kB)\nBuilding wheels for collected packages: mpds-aiida, aiida-crystal-dft, mpds-ml-labs, mpds_client, pycrystal, progressbar, pycodcif\n  Building wheel for mpds-aiida (setup.py): started\n  Building wheel for mpds-aiida (setup.py): finished with status 'done'\n  Created wheel for mpds-aiida: filename=mpds_aiida-0.10.1-py3-none-any.whl size=26528 sha256=0e2a6f4e90f39e33a6326b2751c98940a9939cc3e6a3556716bbdd4628b823a1\n  Stored in directory: /tmp/pip-ephem-wheel-cache-liu3u1u0/wheels/78/53/54/1f9951ffdf9d9f80855f9b673b98a29c0fd5474eda71dda1dd\n  Building wheel for aiida-crystal-dft (pyproject.toml): started\n  Building wheel for aiida-crystal-dft (pyproject.toml): finished with status 'done'\n  Created wheel for aiida-crystal-dft: filename=aiida_crystal_dft-0.9.1-py3-none-any.whl size=615109 sha256=a593e3645f0af4f057dc7e5feee6c57221f560c544b74195a093d03b051e5b68\n  Stored in directory: /tmp/pip-ephem-wheel-cache-liu3u1u0/wheels/a3/99/76/53856ab4749a880e808f4f2a1fa0f13576e6bef968abb3337c\n  Building wheel for mpds-ml-labs (setup.py): started\n  Building wheel for mpds-ml-labs (setup.py): finished with status 'done'\n  Created wheel for mpds-ml-labs: filename=mpds_ml_labs-0.0.7-py3-none-any.whl size=36850 sha256=ae165c43db1c772a0a645a64b93e26c0f30f2c175fc76125bd6a16d00d4dbe44\n  Stored in directory: /tmp/pip-ephem-wheel-cache-liu3u1u0/wheels/4c/ed/d3/e721adc8f33103079f4f5fae1381741cd3f5a08d09cf18e59b\n  Building wheel for mpds_client (setup.py): started\n  Building wheel for mpds_client (setup.py): finished with status 'done'\n  Created wheel for mpds_client: filename=mpds_client-0.24-py3-none-any.whl size=9969 sha256=a544ef102fea4ae53a1b2a6f13f8f5e04926137999049af2cfa870bee70df345\n  Stored in directory: /home/aiida/.cache/pip/wheels/ac/3b/35/86c36e5a6497d198803077160a5ef38ef680c3a20750d01cdb\n  Building wheel for pycrystal (setup.py): started\n  Building wheel for pycrystal (setup.py): finished with status 'done'\n  Created wheel for pycrystal: filename=pycrystal-1.0.16-py3-none-any.whl size=27471 sha256=88cfd15e56b4d8c8ebae92515effe3fff67072edda5f6c988bbe9a6666a51a10\n  Stored in directory: /home/aiida/.cache/pip/wheels/bb/b8/bd/373832f07727578241660773b3994e17776c78ca6a241f3362\n  Building wheel for progressbar (setup.py): started\n  Building wheel for progressbar (setup.py): finished with status 'done'\n  Created wheel for progressbar: filename=progressbar-2.5-py3-none-any.whl size=12067 sha256=e11c6c1ee5fe482c54742da8a68de2c72d5e301d2a7c68875898db862555bde8\n  Stored in directory: /home/aiida/.cache/pip/wheels/cd/17/e5/765d1a3112ff3978f70223502f6047e06c43a24d7c5f8ff95b\n  Building wheel for pycodcif (setup.py): started\n  Building wheel for pycodcif (setup.py): finished with status 'error'\n  error: subprocess-exited-with-error\n  \n  \u00d7 python setup.py bdist_wheel did not run successfully.\n  \u2502 exit code: 1\n  \u2570\u2500> [8 lines of output]\n      running bdist_wheel\n      running build\n      running build_py\n      running build_ext\n      building 'pycodcif._pycodcif' extension\n      swigging pycodcif.i to pycodcif_wrap.c\n      swig -python -o pycodcif_wrap.c pycodcif.i\n      error: command 'swig' failed: No such file or directory\n      [end of output]\n  \n  note: This error originates from a subprocess, and is likely not a problem with pip.\n  ERROR: Failed building wheel for pycodcif\n  Running setup.py clean for pycodcif\nSuccessfully built mpds-aiida aiida-crystal-dft mpds-ml-labs mpds_client pycrystal progressbar\nFailed to build pycodcif\nERROR: Could not build wheels for pycodcif, which is required to install pyproject.toml-based projects\n
" + "E001: Failed to install plugin aiida-mpds
Collecting git+https://github.com/mpds-io/mpds-aiida\n  Cloning https://github.com/mpds-io/mpds-aiida to /tmp/pip-req-build-sw4od4h5\n  Running command git clone --filter=blob:none --quiet https://github.com/mpds-io/mpds-aiida /tmp/pip-req-build-sw4od4h5\n  Resolved https://github.com/mpds-io/mpds-aiida to commit b4ab96db1472969e89159b2b2d76a9836b99c94a\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs (from mpds-aiida==0.10.1)\n  Cloning https://github.com/mpds-io/mpds-ml-labs to /tmp/pip-install-2y_nm2nh/mpds-ml-labs_601d7a0c154742ebb1990bde07b054db\n  Running command git clone --filter=blob:none --quiet https://github.com/mpds-io/mpds-ml-labs /tmp/pip-install-2y_nm2nh/mpds-ml-labs_601d7a0c154742ebb1990bde07b054db\n  Resolved https://github.com/mpds-io/mpds-ml-labs to commit 46b15d3fb27a1ca3389d9fbe44b9061740291bc7\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft (from mpds-aiida==0.10.1)\n  Cloning https://github.com/tilde-lab/aiida-crystal-dft to /tmp/pip-install-2y_nm2nh/aiida-crystal-dft_e25f162197a5436aa2c2bb22378b3727\n  Running command git clone --filter=blob:none --quiet https://github.com/tilde-lab/aiida-crystal-dft /tmp/pip-install-2y_nm2nh/aiida-crystal-dft_e25f162197a5436aa2c2bb22378b3727\n  Resolved https://github.com/tilde-lab/aiida-crystal-dft to commit dfa54170debe4be80c9fc8983e826af46afe5916\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nCollecting reentry>=1.3.1 (from mpds-aiida==0.10.1)\n  Using cached reentry-1.3.3-py3-none-any.whl.metadata (10 kB)\nRequirement already satisfied: six>=1.16.0 in /opt/conda/lib/python3.10/site-packages (from mpds-aiida==0.10.1) (1.16.0)\nRequirement already satisfied: numpy>=1.17.5 in /opt/conda/lib/python3.10/site-packages (from mpds-aiida==0.10.1) (1.26.4)\nCollecting ase>=3.19 (from mpds-aiida==0.10.1)\n  Downloading ase-3.23.0-py3-none-any.whl.metadata (3.8 kB)\nCollecting yascheduler>=1.0.12 (from mpds-aiida==0.10.1)\n  Downloading yascheduler-1.2.0-py3-none-any.whl.metadata (12 kB)\nCollecting scipy>=1.6.0 (from ase>=3.19->mpds-aiida==0.10.1)\n  Downloading scipy-1.14.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (60 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 60.8/60.8 kB 4.3 MB/s eta 0:00:00\nCollecting matplotlib>=3.3.4 (from ase>=3.19->mpds-aiida==0.10.1)\n  Downloading matplotlib-3.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (11 kB)\nRequirement already satisfied: setuptools>=36.2 in /opt/conda/lib/python3.10/site-packages (from reentry>=1.3.1->mpds-aiida==0.10.1) (70.0.0)\nRequirement already satisfied: click in /opt/conda/lib/python3.10/site-packages (from reentry>=1.3.1->mpds-aiida==0.10.1) (8.1.7)\nCollecting aiohttp~=3.8 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading aiohttp-3.11.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)\nCollecting asyncssh~=2.11 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading asyncssh-2.18.0-py3-none-any.whl.metadata (9.9 kB)\nCollecting asyncstdlib~=3.10 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading asyncstdlib-3.13.0-py3-none-any.whl.metadata (3.7 kB)\nCollecting attrs~=21.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading attrs-21.4.0-py2.py3-none-any.whl.metadata (9.8 kB)\nCollecting azure-identity~=1.10.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_identity-1.10.0-py3-none-any.whl.metadata (54 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 54.5/54.5 kB 12.9 MB/s eta 0:00:00\nCollecting azure-mgmt-compute~=27.2.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_mgmt_compute-27.2.0-py3-none-any.whl.metadata (53 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.1/53.1 kB 12.5 MB/s eta 0:00:00\nCollecting azure-mgmt-network~=20.0.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_mgmt_network-20.0.0-py3-none-any.whl.metadata (63 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 63.9/63.9 kB 21.3 MB/s eta 0:00:00\nCollecting backoff~=2.1.2 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading backoff-2.1.2-py3-none-any.whl.metadata (13 kB)\nCollecting hcloud~=1.17 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading hcloud-1.35.0-py3-none-any.whl.metadata (4.9 kB)\nCollecting pg8000~=1.19 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading pg8000-1.31.2-py3-none-any.whl.metadata (74 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 74.7/74.7 kB 18.0 MB/s eta 0:00:00\nCollecting python-daemon~=2.3 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading python_daemon-2.3.2-py3-none-any.whl.metadata (2.0 kB)\nRequirement already satisfied: typing-extensions>=4.2.0 in /opt/conda/lib/python3.10/site-packages (from yascheduler>=1.0.12->mpds-aiida==0.10.1) (4.12.1)\nCollecting upcloud_api~=2.0 (from yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading upcloud_api-2.6.0-py3-none-any.whl.metadata (7.8 kB)\nRequirement already satisfied: aiida-core>=2.0.2 in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.5.2)\nCollecting pycrystal>=1.0.10 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pycrystal-1.0.16.tar.gz (28 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting mpds_client>=0.24 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading mpds_client-0.24.tar.gz (9.7 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting pyparsing>2.3.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pyparsing-3.2.0-py3-none-any.whl.metadata (5.0 kB)\nCollecting spglib==2.0.2 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading spglib-2.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (1.3 kB)\nRequirement already satisfied: jinja2>=2.10 in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.1.4)\nCollecting jsonschema>=3.0.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading jsonschema-4.23.0-py3-none-any.whl.metadata (7.9 kB)\nRequirement already satisfied: packaging in /opt/conda/lib/python3.10/site-packages (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (24.0)\nCollecting pycodcif (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading pycodcif-3.0.1.tar.gz (79 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 79.6/79.6 kB 21.5 MB/s eta 0:00:00\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nCollecting scikit-learn (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading scikit_learn-1.5.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (13 kB)\nCollecting imblearn (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading imblearn-0.0-py2.py3-none-any.whl.metadata (355 bytes)\nCollecting progressbar (from mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading progressbar-2.5.tar.gz (10 kB)\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nRequirement already satisfied: alembic~=1.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.13.1)\nRequirement already satisfied: archive-path~=0.4.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.4.2)\nRequirement already satisfied: aio-pika~=6.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.8.1)\nRequirement already satisfied: circus~=0.18.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.18.0)\nRequirement already satisfied: click-spinner~=0.1.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.1.10)\nRequirement already satisfied: disk-objectstore~=1.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.1.0)\nRequirement already satisfied: docstring-parser in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.16)\nRequirement already satisfied: graphviz~=0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.20.3)\nRequirement already satisfied: ipython>=7 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (8.25.0)\nRequirement already satisfied: jedi<0.19 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.18.2)\nRequirement already satisfied: kiwipy~=0.7.7 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.8)\nRequirement already satisfied: importlib-metadata~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.11.0)\nRequirement already satisfied: paramiko>=2.7.2,~=2.7 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.12.0)\nRequirement already satisfied: plumpy~=0.21.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.21.10)\nRequirement already satisfied: pgsu~=0.2.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.4)\nRequirement already satisfied: psutil~=5.6 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.9.8)\nRequirement already satisfied: psycopg2-binary~=2.8 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.9.9)\nRequirement already satisfied: pydantic~=2.4 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.7.3)\nRequirement already satisfied: pytz~=2021.1 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2021.3)\nRequirement already satisfied: pyyaml~=6.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.0.1)\nRequirement already satisfied: requests~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.32.3)\nRequirement already satisfied: sqlalchemy~=2.0 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.0.30)\nRequirement already satisfied: tabulate~=0.8.5 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.8.10)\nRequirement already satisfied: tqdm~=4.45 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.66.4)\nRequirement already satisfied: upf_to_json~=0.9.2 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.9.5)\nRequirement already satisfied: wrapt~=1.11 in /opt/conda/lib/python3.10/site-packages (from aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.16.0)\nCollecting aiohappyeyeballs>=2.3.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading aiohappyeyeballs-2.4.3-py3-none-any.whl.metadata (6.1 kB)\nCollecting aiosignal>=1.1.2 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading aiosignal-1.3.1-py3-none-any.whl.metadata (4.0 kB)\nCollecting async-timeout<6.0,>=4.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading async_timeout-5.0.1-py3-none-any.whl.metadata (5.1 kB)\nCollecting frozenlist>=1.1.1 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (13 kB)\nRequirement already satisfied: multidict<7.0,>=4.5 in /opt/conda/lib/python3.10/site-packages (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1) (6.0.5)\nCollecting propcache>=0.2.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading propcache-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (7.7 kB)\nCollecting yarl<2.0,>=1.17.0 (from aiohttp~=3.8->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading yarl-1.18.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (67 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 67.7/67.7 kB 16.5 MB/s eta 0:00:00\nRequirement already satisfied: cryptography>=39.0 in /opt/conda/lib/python3.10/site-packages (from asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (42.0.8)\nCollecting azure-core<2.0.0,>=1.11.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_core-1.32.0-py3-none-any.whl.metadata (39 kB)\nCollecting msal<2.0.0,>=1.12.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading msal-1.31.1-py3-none-any.whl.metadata (11 kB)\nCollecting msal-extensions<2.0.0,>=0.3.0 (from azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading msal_extensions-1.2.0-py3-none-any.whl.metadata (7.6 kB)\nCollecting msrest>=0.6.21 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading msrest-0.7.1-py3-none-any.whl.metadata (21 kB)\nCollecting azure-common~=1.1 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_common-1.1.28-py2.py3-none-any.whl.metadata (5.0 kB)\nCollecting azure-mgmt-core<2.0.0,>=1.3.1 (from azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading azure_mgmt_core-1.5.0-py3-none-any.whl.metadata (4.3 kB)\nCollecting python-dateutil>=2.7.5 (from hcloud~=1.17->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl.metadata (8.4 kB)\nRequirement already satisfied: MarkupSafe>=2.0 in /opt/conda/lib/python3.10/site-packages (from jinja2>=2.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.1.5)\nINFO: pip is looking at multiple versions of jsonschema to determine which version is compatible with other requirements. This could take a while.\nCollecting jsonschema>=3.0.1 (from aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading jsonschema-4.22.0-py3-none-any.whl.metadata (8.2 kB)\n  Downloading jsonschema-4.21.1-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.21.0-py3-none-any.whl.metadata (8.0 kB)\n  Downloading jsonschema-4.20.0-py3-none-any.whl.metadata (8.1 kB)\n  Downloading jsonschema-4.19.2-py3-none-any.whl.metadata (7.9 kB)\n  Downloading jsonschema-4.19.1-py3-none-any.whl.metadata (7.9 kB)\n  Downloading jsonschema-4.19.0-py3-none-any.whl.metadata (8.2 kB)\nINFO: pip is still looking at multiple versions of jsonschema to determine which version is compatible with other requirements. This could take a while.\n  Downloading jsonschema-4.18.6-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.18.5-py3-none-any.whl.metadata (7.7 kB)\n  Downloading jsonschema-4.18.4-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.18.3-py3-none-any.whl.metadata (7.9 kB)\n  Downloading jsonschema-4.18.2-py3-none-any.whl.metadata (7.8 kB)\nINFO: This is taking longer than usual. You might need to provide the dependency resolver with stricter constraints to reduce runtime. See https://pip.pypa.io/warnings/backtracking for guidance. If you want to abort this run, press Ctrl + C.\n  Downloading jsonschema-4.18.1-py3-none-any.whl.metadata (7.8 kB)\n  Downloading jsonschema-4.18.0-py3-none-any.whl.metadata (10 kB)\n  Downloading jsonschema-4.17.3-py3-none-any.whl.metadata (7.9 kB)\nCollecting pyrsistent!=0.17.0,!=0.17.1,!=0.17.2,>=0.14.0 (from jsonschema>=3.0.1->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pyrsistent-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (27 kB)\nCollecting contourpy>=1.0.1 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading contourpy-1.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (5.4 kB)\nCollecting cycler>=0.10 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading cycler-0.12.1-py3-none-any.whl.metadata (3.8 kB)\nCollecting fonttools>=4.22.0 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading fonttools-4.55.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (164 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 164.5/164.5 kB 13.7 MB/s eta 0:00:00\nCollecting kiwisolver>=1.3.1 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading kiwisolver-1.4.7-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl.metadata (6.3 kB)\nCollecting pillow>=8 (from matplotlib>=3.3.4->ase>=3.19->mpds-aiida==0.10.1)\n  Downloading pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl.metadata (9.1 kB)\nCollecting httplib2 (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading httplib2-0.22.0-py3-none-any.whl.metadata (2.6 kB)\nCollecting ujson (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading ujson-5.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (9.3 kB)\nCollecting pandas (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl.metadata (89 kB)\n     \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 89.9/89.9 kB 29.1 MB/s eta 0:00:00\nCollecting jmespath (from mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading jmespath-1.0.1-py3-none-any.whl.metadata (7.6 kB)\nCollecting scramp>=1.4.5 (from pg8000~=1.19->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading scramp-1.4.5-py3-none-any.whl.metadata (19 kB)\nCollecting bs4 (from pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading bs4-0.0.2-py2.py3-none-any.whl.metadata (411 bytes)\nCollecting docutils (from python-daemon~=2.3->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading docutils-0.21.2-py3-none-any.whl.metadata (2.8 kB)\nCollecting lockfile>=0.10 (from python-daemon~=2.3->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading lockfile-0.12.2-py2.py3-none-any.whl.metadata (2.4 kB)\nCollecting imbalanced-learn (from imblearn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading imbalanced_learn-0.12.4-py3-none-any.whl.metadata (8.3 kB)\nCollecting joblib>=1.2.0 (from scikit-learn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading joblib-1.4.2-py3-none-any.whl.metadata (5.4 kB)\nCollecting threadpoolctl>=3.1.0 (from scikit-learn->mpds-ml-labs@ git+https://github.com/mpds-io/mpds-ml-labs->mpds-aiida==0.10.1)\n  Downloading threadpoolctl-3.5.0-py3-none-any.whl.metadata (13 kB)\nRequirement already satisfied: aiormq<4,>=3.2.3 in /opt/conda/lib/python3.10/site-packages (from aio-pika~=6.6->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.3.1)\nRequirement already satisfied: Mako in /opt/conda/lib/python3.10/site-packages (from alembic~=1.2->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.3.5)\nRequirement already satisfied: pyzmq>=17.0 in /opt/conda/lib/python3.10/site-packages (from circus~=0.18.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (26.0.3)\nRequirement already satisfied: tornado>=5.0.2 in /opt/conda/lib/python3.10/site-packages (from circus~=0.18.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (6.4)\nRequirement already satisfied: cffi>=1.12 in /opt/conda/lib/python3.10/site-packages (from cryptography>=39.0->asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (1.16.0)\nRequirement already satisfied: zipp>=0.5 in /opt/conda/lib/python3.10/site-packages (from importlib-metadata~=6.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.19.2)\nRequirement already satisfied: decorator in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.1.1)\nRequirement already satisfied: matplotlib-inline in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.1.7)\nRequirement already satisfied: prompt-toolkit<3.1.0,>=3.0.41 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.0.46)\nRequirement already satisfied: pygments>=2.4.0 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.18.0)\nRequirement already satisfied: stack-data in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.6.3)\nRequirement already satisfied: traitlets>=5.13.0 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (5.14.3)\nRequirement already satisfied: exceptiongroup in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.2.1)\nRequirement already satisfied: pexpect>4.3 in /opt/conda/lib/python3.10/site-packages (from ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.9.0)\nRequirement already satisfied: parso<0.9.0,>=0.8.0 in /opt/conda/lib/python3.10/site-packages (from jedi<0.19->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.8.4)\nRequirement already satisfied: shortuuid in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.0.13)\nRequirement already satisfied: async-generator in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.10)\nRequirement already satisfied: pytray<0.4.0,>=0.2.2 in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.3.4)\nRequirement already satisfied: deprecation in /opt/conda/lib/python3.10/site-packages (from kiwipy~=0.7.7->kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.1.0)\nRequirement already satisfied: pamqp~=2.0 in /opt/conda/lib/python3.10/site-packages (from kiwipy[rmq]~=0.7.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.3.0)\nCollecting PyJWT<3,>=1.0.0 (from PyJWT[crypto]<3,>=1.0.0->msal<2.0.0,>=1.12.0->azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading PyJWT-2.10.0-py3-none-any.whl.metadata (4.0 kB)\nCollecting portalocker<3,>=1.4 (from msal-extensions<2.0.0,>=0.3.0->azure-identity~=1.10.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading portalocker-2.10.1-py3-none-any.whl.metadata (8.5 kB)\nRequirement already satisfied: certifi>=2017.4.17 in /opt/conda/lib/python3.10/site-packages (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1) (2024.2.2)\nCollecting isodate>=0.6.0 (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading isodate-0.7.2-py3-none-any.whl.metadata (11 kB)\nCollecting requests-oauthlib>=0.5.0 (from msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading requests_oauthlib-2.0.0-py2.py3-none-any.whl.metadata (11 kB)\nRequirement already satisfied: bcrypt>=3.1.3 in /opt/conda/lib/python3.10/site-packages (from paramiko>=2.7.2,~=2.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (4.1.3)\nRequirement already satisfied: pynacl>=1.0.1 in /opt/conda/lib/python3.10/site-packages (from paramiko>=2.7.2,~=2.7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.5.0)\nRequirement already satisfied: nest_asyncio~=1.5 in /opt/conda/lib/python3.10/site-packages (from plumpy~=0.21.6->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (1.6.0)\nRequirement already satisfied: annotated-types>=0.4.0 in /opt/conda/lib/python3.10/site-packages (from pydantic~=2.4->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.0)\nRequirement already satisfied: pydantic-core==2.18.4 in /opt/conda/lib/python3.10/site-packages (from pydantic~=2.4->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.18.4)\nRequirement already satisfied: charset-normalizer<4,>=2 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.3.2)\nRequirement already satisfied: idna<4,>=2.5 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.7)\nRequirement already satisfied: urllib3<3,>=1.21.1 in /opt/conda/lib/python3.10/site-packages (from requests~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.2.1)\nCollecting asn1crypto>=1.5.1 (from scramp>=1.4.5->pg8000~=1.19->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading asn1crypto-1.5.1-py2.py3-none-any.whl.metadata (13 kB)\nRequirement already satisfied: greenlet!=0.4.17 in /opt/conda/lib/python3.10/site-packages (from sqlalchemy~=2.0->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (3.0.3)\nCollecting beautifulsoup4 (from bs4->pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading beautifulsoup4-4.12.3-py3-none-any.whl.metadata (3.8 kB)\nCollecting tzdata>=2022.7 (from pandas->mpds_client>=0.24->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading tzdata-2024.2-py2.py3-none-any.whl.metadata (1.4 kB)\nRequirement already satisfied: pycparser in /opt/conda/lib/python3.10/site-packages (from cffi>=1.12->cryptography>=39.0->asyncssh~=2.11->yascheduler>=1.0.12->mpds-aiida==0.10.1) (2.22)\nRequirement already satisfied: ptyprocess>=0.5 in /opt/conda/lib/python3.10/site-packages (from pexpect>4.3->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.7.0)\nRequirement already satisfied: wcwidth in /opt/conda/lib/python3.10/site-packages (from prompt-toolkit<3.1.0,>=3.0.41->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.13)\nCollecting oauthlib>=3.0.0 (from requests-oauthlib>=0.5.0->msrest>=0.6.21->azure-mgmt-compute~=27.2.0->yascheduler>=1.0.12->mpds-aiida==0.10.1)\n  Downloading oauthlib-3.2.2-py3-none-any.whl.metadata (7.5 kB)\nCollecting soupsieve>1.2 (from beautifulsoup4->bs4->pycrystal>=1.0.10->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1)\n  Downloading soupsieve-2.6-py3-none-any.whl.metadata (4.6 kB)\nRequirement already satisfied: executing>=1.2.0 in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.0.1)\nRequirement already satisfied: asttokens>=2.1.0 in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (2.4.1)\nRequirement already satisfied: pure-eval in /opt/conda/lib/python3.10/site-packages (from stack-data->ipython>=7->aiida-core>=2.0.2->aiida-crystal-dft@ git+https://github.com/tilde-lab/aiida-crystal-dft->mpds-aiida==0.10.1) (0.2.2)\nDownloading ase-3.23.0-py3-none-any.whl (2.9 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 2.9/2.9 MB 74.6 MB/s eta 0:00:00\nUsing cached reentry-1.3.3-py3-none-any.whl (17 kB)\nDownloading yascheduler-1.2.0-py3-none-any.whl (67 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 67.9/67.9 kB 21.7 MB/s eta 0:00:00\nDownloading spglib-2.0.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (515 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 515.3/515.3 kB 76.2 MB/s eta 0:00:00\nDownloading aiohttp-3.11.7-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (1.6 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.6/1.6 MB 105.2 MB/s eta 0:00:00\nDownloading asyncssh-2.18.0-py3-none-any.whl (367 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 367.6/367.6 kB 63.8 MB/s eta 0:00:00\nDownloading asyncstdlib-3.13.0-py3-none-any.whl (43 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 43.9/43.9 kB 10.0 MB/s eta 0:00:00\nDownloading attrs-21.4.0-py2.py3-none-any.whl (60 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 60.6/60.6 kB 18.9 MB/s eta 0:00:00\nDownloading azure_identity-1.10.0-py3-none-any.whl (134 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 134.1/134.1 kB 36.8 MB/s eta 0:00:00\nDownloading azure_mgmt_compute-27.2.0-py3-none-any.whl (5.7 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 5.7/5.7 MB 47.0 MB/s eta 0:00:00\nDownloading azure_mgmt_network-20.0.0-py3-none-any.whl (8.5 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 8.5/8.5 MB 115.1 MB/s eta 0:00:00\nDownloading backoff-2.1.2-py3-none-any.whl (14 kB)\nDownloading hcloud-1.35.0-py3-none-any.whl (85 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 85.4/85.4 kB 21.5 MB/s eta 0:00:00\nDownloading jsonschema-4.17.3-py3-none-any.whl (90 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 90.4/90.4 kB 27.0 MB/s eta 0:00:00\nDownloading matplotlib-3.9.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (8.3 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 8.3/8.3 MB 117.1 MB/s eta 0:00:00\nDownloading pg8000-1.31.2-py3-none-any.whl (54 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 54.5/54.5 kB 16.4 MB/s eta 0:00:00\nDownloading pyparsing-3.2.0-py3-none-any.whl (106 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 106.9/106.9 kB 31.0 MB/s eta 0:00:00\nDownloading python_daemon-2.3.2-py3-none-any.whl (34 kB)\nDownloading scipy-1.14.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (41.2 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 41.2/41.2 MB 69.5 MB/s eta 0:00:00\nDownloading upcloud_api-2.6.0-py3-none-any.whl (37 kB)\nDownloading imblearn-0.0-py2.py3-none-any.whl (1.9 kB)\nDownloading scikit_learn-1.5.2-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.3 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 13.3/13.3 MB 118.0 MB/s eta 0:00:00\nDownloading aiohappyeyeballs-2.4.3-py3-none-any.whl (14 kB)\nDownloading aiosignal-1.3.1-py3-none-any.whl (7.6 kB)\nDownloading async_timeout-5.0.1-py3-none-any.whl (6.2 kB)\nDownloading azure_common-1.1.28-py2.py3-none-any.whl (14 kB)\nDownloading azure_core-1.32.0-py3-none-any.whl (198 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 198.9/198.9 kB 47.3 MB/s eta 0:00:00\nDownloading azure_mgmt_core-1.5.0-py3-none-any.whl (30 kB)\nDownloading contourpy-1.3.1-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (324 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 325.0/325.0 kB 65.3 MB/s eta 0:00:00\nDownloading cycler-0.12.1-py3-none-any.whl (8.3 kB)\nDownloading fonttools-4.55.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (4.6 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 4.6/4.6 MB 113.3 MB/s eta 0:00:00\nDownloading frozenlist-1.5.0-cp310-cp310-manylinux_2_5_x86_64.manylinux1_x86_64.manylinux_2_17_x86_64.manylinux2014_x86_64.whl (241 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 241.9/241.9 kB 51.4 MB/s eta 0:00:00\nDownloading joblib-1.4.2-py3-none-any.whl (301 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 301.8/301.8 kB 62.3 MB/s eta 0:00:00\nDownloading kiwisolver-1.4.7-cp310-cp310-manylinux_2_12_x86_64.manylinux2010_x86_64.whl (1.6 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 1.6/1.6 MB 102.4 MB/s eta 0:00:00\nDownloading lockfile-0.12.2-py2.py3-none-any.whl (13 kB)\nDownloading msal-1.31.1-py3-none-any.whl (113 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 113.2/113.2 kB 30.9 MB/s eta 0:00:00\nDownloading msal_extensions-1.2.0-py3-none-any.whl (19 kB)\nDownloading msrest-0.7.1-py3-none-any.whl (85 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 85.4/85.4 kB 19.9 MB/s eta 0:00:00\nDownloading pillow-11.0.0-cp310-cp310-manylinux_2_28_x86_64.whl (4.4 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 4.4/4.4 MB 118.7 MB/s eta 0:00:00\nDownloading propcache-0.2.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (208 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 208.9/208.9 kB 36.7 MB/s eta 0:00:00\nDownloading pyrsistent-0.20.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (117 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 117.7/117.7 kB 27.4 MB/s eta 0:00:00\nDownloading python_dateutil-2.9.0.post0-py2.py3-none-any.whl (229 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 229.9/229.9 kB 42.9 MB/s eta 0:00:00\nDownloading scramp-1.4.5-py3-none-any.whl (12 kB)\nDownloading threadpoolctl-3.5.0-py3-none-any.whl (18 kB)\nDownloading yarl-1.18.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (319 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 319.4/319.4 kB 67.3 MB/s eta 0:00:00\nDownloading bs4-0.0.2-py2.py3-none-any.whl (1.2 kB)\nDownloading docutils-0.21.2-py3-none-any.whl (587 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 587.4/587.4 kB 43.1 MB/s eta 0:00:00\nDownloading httplib2-0.22.0-py3-none-any.whl (96 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 96.9/96.9 kB 26.1 MB/s eta 0:00:00\nDownloading imbalanced_learn-0.12.4-py3-none-any.whl (258 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 258.3/258.3 kB 47.1 MB/s eta 0:00:00\nDownloading jmespath-1.0.1-py3-none-any.whl (20 kB)\nDownloading pandas-2.2.3-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (13.1 MB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 13.1/13.1 MB 118.5 MB/s eta 0:00:00\nDownloading ujson-5.10.0-cp310-cp310-manylinux_2_17_x86_64.manylinux2014_x86_64.whl (53 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 53.6/53.6 kB 15.2 MB/s eta 0:00:00\nDownloading asn1crypto-1.5.1-py2.py3-none-any.whl (105 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 105.0/105.0 kB 31.3 MB/s eta 0:00:00\nDownloading isodate-0.7.2-py3-none-any.whl (22 kB)\nDownloading portalocker-2.10.1-py3-none-any.whl (18 kB)\nDownloading PyJWT-2.10.0-py3-none-any.whl (23 kB)\nDownloading requests_oauthlib-2.0.0-py2.py3-none-any.whl (24 kB)\nDownloading tzdata-2024.2-py2.py3-none-any.whl (346 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 346.6/346.6 kB 62.6 MB/s eta 0:00:00\nDownloading beautifulsoup4-4.12.3-py3-none-any.whl (147 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 147.9/147.9 kB 38.3 MB/s eta 0:00:00\nDownloading oauthlib-3.2.2-py3-none-any.whl (151 kB)\n   \u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501\u2501 151.7/151.7 kB 42.8 MB/s eta 0:00:00\nDownloading soupsieve-2.6-py3-none-any.whl (36 kB)\nBuilding wheels for collected packages: mpds-aiida, aiida-crystal-dft, mpds-ml-labs, mpds_client, pycrystal, progressbar, pycodcif\n  Building wheel for mpds-aiida (setup.py): started\n  Building wheel for mpds-aiida (setup.py): finished with status 'done'\n  Created wheel for mpds-aiida: filename=mpds_aiida-0.10.1-py3-none-any.whl size=26528 sha256=425ccfd63a745288db9c902e668f77529951df66db7a8f44321478923363b091\n  Stored in directory: /tmp/pip-ephem-wheel-cache-erc7izfh/wheels/78/53/54/1f9951ffdf9d9f80855f9b673b98a29c0fd5474eda71dda1dd\n  Building wheel for aiida-crystal-dft (pyproject.toml): started\n  Building wheel for aiida-crystal-dft (pyproject.toml): finished with status 'done'\n  Created wheel for aiida-crystal-dft: filename=aiida_crystal_dft-0.9.1-py3-none-any.whl size=615109 sha256=823d3da78287542b4ac60ffca9a7448e9e38fc571591562ced16f8690daf6846\n  Stored in directory: /tmp/pip-ephem-wheel-cache-erc7izfh/wheels/a3/99/76/53856ab4749a880e808f4f2a1fa0f13576e6bef968abb3337c\n  Building wheel for mpds-ml-labs (setup.py): started\n  Building wheel for mpds-ml-labs (setup.py): finished with status 'done'\n  Created wheel for mpds-ml-labs: filename=mpds_ml_labs-0.0.7-py3-none-any.whl size=36850 sha256=6f681fe74e2df6e644b1d13721fa2dc241f6c2616eafe0438641dbfca9a82b5a\n  Stored in directory: /tmp/pip-ephem-wheel-cache-erc7izfh/wheels/4c/ed/d3/e721adc8f33103079f4f5fae1381741cd3f5a08d09cf18e59b\n  Building wheel for mpds_client (setup.py): started\n  Building wheel for mpds_client (setup.py): finished with status 'done'\n  Created wheel for mpds_client: filename=mpds_client-0.24-py3-none-any.whl size=9969 sha256=66dd9efb0add0ec83f4c18ffd23e127450f6c8a03ca33e7b154288e51dda88a3\n  Stored in directory: /home/aiida/.cache/pip/wheels/ac/3b/35/86c36e5a6497d198803077160a5ef38ef680c3a20750d01cdb\n  Building wheel for pycrystal (setup.py): started\n  Building wheel for pycrystal (setup.py): finished with status 'done'\n  Created wheel for pycrystal: filename=pycrystal-1.0.16-py3-none-any.whl size=27471 sha256=ef41fccfee9ffab7a3812b4aca11fa0b34b09243ee8b1bd553a5533b7a820036\n  Stored in directory: /home/aiida/.cache/pip/wheels/bb/b8/bd/373832f07727578241660773b3994e17776c78ca6a241f3362\n  Building wheel for progressbar (setup.py): started\n  Building wheel for progressbar (setup.py): finished with status 'done'\n  Created wheel for progressbar: filename=progressbar-2.5-py3-none-any.whl size=12067 sha256=2df91d6baab15b8717135f02873c6d5691312ddc9800fef4d866dc9330e5add3\n  Stored in directory: /home/aiida/.cache/pip/wheels/cd/17/e5/765d1a3112ff3978f70223502f6047e06c43a24d7c5f8ff95b\n  Building wheel for pycodcif (setup.py): started\n  Building wheel for pycodcif (setup.py): finished with status 'error'\n  error: subprocess-exited-with-error\n  \n  \u00d7 python setup.py bdist_wheel did not run successfully.\n  \u2502 exit code: 1\n  \u2570\u2500> [8 lines of output]\n      running bdist_wheel\n      running build\n      running build_py\n      running build_ext\n      building 'pycodcif._pycodcif' extension\n      swigging pycodcif.i to pycodcif_wrap.c\n      swig -python -o pycodcif_wrap.c pycodcif.i\n      error: command 'swig' failed: No such file or directory\n      [end of output]\n  \n  note: This error originates from a subprocess, and is likely not a problem with pip.\n  ERROR: Failed building wheel for pycodcif\n  Running setup.py clean for pycodcif\nSuccessfully built mpds-aiida aiida-crystal-dft mpds-ml-labs mpds_client pycrystal progressbar\nFailed to build pycodcif\nERROR: Could not build wheels for pycodcif, which is required to install pyproject.toml-based projects\n
" ], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead.", @@ -21723,7 +21723,7 @@ } } }, - "commits_count": 85, + "commits_count": 84, "development_status": "stable", "errors": [], "warnings": [], @@ -27034,7 +27034,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-psi4
Collecting git+https://github.com/ltalirz/aiida-psi4\n  Cloning https://github.com/ltalirz/aiida-psi4 to /tmp/pip-req-build-tqqjjp5t\n  Running command git clone --filter=blob:none --quiet https://github.com/ltalirz/aiida-psi4 /tmp/pip-req-build-tqqjjp5t\n  Resolved https://github.com/ltalirz/aiida-psi4 to commit 637e6b0b29e724a158014269d55d9091c6af48c7\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Installing backend dependencies: started\n  Installing backend dependencies: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-psi4 to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-psi4==0.1.0a0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-psi4 0.1.0a0 depends on aiida-core<2.0.0 and >=1.6.4\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-psi4
Collecting git+https://github.com/ltalirz/aiida-psi4\n  Cloning https://github.com/ltalirz/aiida-psi4 to /tmp/pip-req-build-cfuplcbg\n  Running command git clone --filter=blob:none --quiet https://github.com/ltalirz/aiida-psi4 /tmp/pip-req-build-cfuplcbg\n  Resolved https://github.com/ltalirz/aiida-psi4 to commit 637e6b0b29e724a158014269d55d9091c6af48c7\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Installing backend dependencies: started\n  Installing backend dependencies: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-psi4 to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-psi4==0.1.0a0 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-psi4 0.1.0a0 depends on aiida-core<2.0.0 and >=1.6.4\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W006: 'development_status' key is deprecated. Use PyPI Trove classifiers in the plugin repository instead." @@ -37224,7 +37224,7 @@ }, "commits_count": 0, "errors": [ - "E001: Failed to install plugin aiida-supercell
Collecting git+https://github.com/pzarabadip/aiida-supercell\n  Cloning https://github.com/pzarabadip/aiida-supercell to /tmp/pip-req-build-j690y9z6\n  Running command git clone --filter=blob:none --quiet https://github.com/pzarabadip/aiida-supercell /tmp/pip-req-build-j690y9z6\n  Resolved https://github.com/pzarabadip/aiida-supercell to commit 24a89f990fcf84584a51f49b53311a5c6db18347\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-supercell to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-supercell==1.0.1 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-supercell 1.0.1 depends on aiida-core<2.0 and >=1.0.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-supercell
Collecting git+https://github.com/pzarabadip/aiida-supercell\n  Cloning https://github.com/pzarabadip/aiida-supercell to /tmp/pip-req-build-4r96dutg\n  Running command git clone --filter=blob:none --quiet https://github.com/pzarabadip/aiida-supercell /tmp/pip-req-build-4r96dutg\n  Resolved https://github.com/pzarabadip/aiida-supercell to commit 24a89f990fcf84584a51f49b53311a5c6db18347\n  Preparing metadata (setup.py): started\n  Preparing metadata (setup.py): finished with status 'done'\nINFO: pip is looking at multiple versions of aiida-supercell to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-supercell==1.0.1 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-supercell 1.0.1 depends on aiida-core<2.0 and >=1.0.0\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [ "W003: Missing classifier 'Framework :: AiiDA'", @@ -44963,7 +44963,7 @@ "commits_count": 0, "development_status": "stable", "errors": [ - "E001: Failed to install plugin aiida-z2pack
Collecting git+https://github.com/AntimoMarrazzo/aiida-z2pack\n  Cloning https://github.com/AntimoMarrazzo/aiida-z2pack to /tmp/pip-req-build-c3vwbzd3\n  Running command git clone --filter=blob:none --quiet https://github.com/AntimoMarrazzo/aiida-z2pack /tmp/pip-req-build-c3vwbzd3\n  Resolved https://github.com/AntimoMarrazzo/aiida-z2pack to commit 01a06de726c8b3eab7d18d1a36b4631ad8216f8b\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nRequirement already satisfied: numpy in /opt/conda/lib/python3.10/site-packages (from aiida-z2pack==3.0.1) (1.26.4)\nCollecting z2pack==2.1.1 (from aiida-z2pack==3.0.1)\n  Downloading z2pack-2.1.1-py3-none-any.whl.metadata (2.3 kB)\nINFO: pip is looking at multiple versions of aiida-z2pack to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-z2pack==3.0.1 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-z2pack 3.0.1 depends on aiida-core<=2.2.2 and >=2.1.2\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" + "E001: Failed to install plugin aiida-z2pack
Collecting git+https://github.com/AntimoMarrazzo/aiida-z2pack\n  Cloning https://github.com/AntimoMarrazzo/aiida-z2pack to /tmp/pip-req-build-4c5_7zw5\n  Running command git clone --filter=blob:none --quiet https://github.com/AntimoMarrazzo/aiida-z2pack /tmp/pip-req-build-4c5_7zw5\n  Resolved https://github.com/AntimoMarrazzo/aiida-z2pack to commit 01a06de726c8b3eab7d18d1a36b4631ad8216f8b\n  Installing build dependencies: started\n  Installing build dependencies: finished with status 'done'\n  Getting requirements to build wheel: started\n  Getting requirements to build wheel: finished with status 'done'\n  Preparing metadata (pyproject.toml): started\n  Preparing metadata (pyproject.toml): finished with status 'done'\nRequirement already satisfied: numpy in /opt/conda/lib/python3.10/site-packages (from aiida-z2pack==3.0.1) (1.26.4)\nCollecting z2pack==2.1.1 (from aiida-z2pack==3.0.1)\n  Downloading z2pack-2.1.1-py3-none-any.whl.metadata (2.3 kB)\nINFO: pip is looking at multiple versions of aiida-z2pack to determine which version is compatible with other requirements. This could take a while.\nERROR: Cannot install aiida-z2pack==3.0.1 because these package versions have conflicting dependencies.\n\nThe conflict is caused by:\n    aiida-z2pack 3.0.1 depends on aiida-core<=2.2.2 and >=2.1.2\n    The user requested (constraint) aiida-core==2.5.2\n\nTo fix this you could try to:\n1. loosen the range of package versions you've specified\n2. remove package versions to allow pip attempt to solve the dependency conflict\n\nERROR: ResolutionImpossible: for help visit https://pip.pypa.io/en/latest/topics/dependency-resolution/#dealing-with-dependency-conflicts\n
" ], "warnings": [], "summaryinfo": [