text
stringlengths
1
1.05M
fn process_mp3_frame(frame: &[u8]) -> Option<(u32, u32)> { if frame.len() < 4 { return None; } let frame_sync = (frame[0] as u16) << 4 | (frame[1] as u16 >> 4); let mpeg_version = (frame[1] >> 3) & 0b11; let layer = (frame[1] >> 1) & 0b11; let protection = frame[1] & 0b1; let bitrate_index = (frame[2] >> 4) & 0b1111; let sample_rate_index = (frame[2] >> 2) & 0b11; let sample_rate = match mpeg_version { 0b00 => match sample_rate_index { 0b00 => 11025, 0b01 => 12000, 0b10 => 8000, _ => return None, }, 0b01 => match sample_rate_index { 0b00 => 22050, 0b01 => 24000, 0b10 => 16000, _ => return None, }, 0b10 => match sample_rate_index { 0b00 => 44100, 0b01 => 48000, 0b10 => 32000, _ => return None, }, _ => return None, }; let bitrate = match (mpeg_version, layer, bitrate_index) { (0b01, 0b11, _) => return None, // Reserved (_, _, 0b0000) => return None, // Free (_, _, 0b1111) => return None, // Bad (0b11, 0b01, _) => return None, // Layer I (0b11, 0b10, _) => return None, // Layer II (0b11, 0b11, _) => return None, // Layer III (_, _, _) => { let bitrates = [ [0, 32, 64, 96, 128, 160, 192, 224, 256, 288, 320, 352, 384, 416, 448], [0, 32, 48, 56, 64, 80, 96, 112, 128, 160, 192, 224, 256, 320, 384], [0, 32, 40, 48, 56, 64, 80, 96, 112, 128, 160, 192, 224, 256, 320], ]; bitrates[mpeg_version as usize - 1][(bitrate_index - 1) as usize] } }; Some((sample_rate, bitrate)) }
#!/usr/bin/env bash source streamlit/bin/activate streamlit run expenses.py data/config.ini
<filename>libs/desktop/shared/ui/src/lib/antd/Picker/DatePicker/index.ts import { Dayjs } from 'dayjs' import dayjsGenerateConfig from 'rc-picker/es/generate/dayjs' import generatePicker from 'antd/es/date-picker/generatePicker' import 'antd/es/date-picker/style/css' const DatePicker = generatePicker<Dayjs>(dayjsGenerateConfig) export { DatePicker }
using System; public class GeometricShape { public string Name { get; set; } public string Color { get; set; } public double Area { get; set; } public override string ToString() { return $"Name: {Name}, Color: {Color}, Area: {Area}"; } } class Program { static void Main() { GeometricShape circle = new GeometricShape { Name = "Circle", Color = "Red", Area = 25.0 }; GeometricShape rectangle = new GeometricShape { Name = "Rectangle", Color = "Blue", Area = 40.5 }; Console.WriteLine("Circle: " + circle); Console.WriteLine("Rectangle: " + rectangle); } }
#!/bin/sh IMAGE_NAME=oci-oke-cli IMAGE_TAG=0.2.0 REPOSITORY_ID=$IMAGE_NAME:$IMAGE_TAG if [ -z $1 ]; then REPOSITORY_ID=$IMAGE_NAME:$IMAGE_TAG else REPOSITORY_ID=$1/$IMAGE_NAME:$IMAGE_TAG fi docker build --build-arg BUILD_DATE=`date -u +”%Y-%m-%dT%H:%M:%SZ”` \ --build-arg VCS_REF=`git rev-parse --short HEAD` \ --build-arg VERSION=$IMAGE_TAG \ -t $REPOSITORY_ID .
package tkohdk.lib.calcstr.checker; import java.util.Arrays; /** * Created by takeoh on 2018/04/19. */ public class OperatorChecker implements OperatorCheckerInterface { /** * 与えられた文字(列)が演算子かどうかを判定する * @param val 演算子かどうかを判定する文字列 * @return boolean */ public boolean isOperator(String val){ String[] op = {"*", "/","+", "-", "×", "÷", "^", "!", "E", "e"}; return Arrays.asList(op).contains(val); } @Override public boolean isBinaryOperator(String val) { String[] op = {"*", "/","+", "-", "×", "÷", "^"}; return Arrays.asList(op).contains(val); } @Override public boolean isUnaryOperator(String val) { // TODO: "E", "e" String[] op = {"!"}; return Arrays.asList(op).contains(val); } //TODO: 単項演算子、二項演算子の判定メソッドを追加 }
#! /bin/sh echo "resetting to 0,0,0,0" rostopic pub -1 /robot/joint1_position_controller/command std_msgs/Float64 "data: 0.0" & rostopic pub -1 /robot/joint2_position_controller/command std_msgs/Float64 "data: 0.0" & rostopic pub -1 /robot/joint3_position_controller/command std_msgs/Float64 "data: 0.0" & rostopic pub -1 /robot/joint4_position_controller/command std_msgs/Float64 "data: 0.0"
#!/bin/bash CATEGORY="$1" ACTION="$2" if [ $CATEGORY == "clean" ]; then if [ $ACTION == "yarn" ]; then echo "Yarn garbage collection started" echo "" echo " - dropping node_modules folder..." rm -rf node_modules echo " Done" echo " - removing .yarn cached files..." rm -rf .yarn/cache .yarn/build-state.yml .yarn/install-state.gz echo " Done" echo " - cleaning yarn cache..." yarn cache clean --all printf " Done\n\n" echo "All done!" elif [ $ACTION == "nuxt" ]; then echo "Nuxtjs garbage collection started" echo "" echo " - dropping .nuxt folder..." rm -rf .nuxt echo " Done" echo " - cleaning the files inside public folder..." rm -rf public/* printf " Done\n\n" echo "All done!" fi fi if [ $CATEGORY == "ci" ]; then echo "Yarn pre-deploy garbage collection started" echo "" bash cli-dev.sh clean nuxt echo "" echo "" bash cli-dev.sh clean yarn echo "" echo "Running yarn install " yarn install --immutable echo "Done" fi
// Copyright © 2018 <NAME> // // Permission is hereby granted, free of charge, to any person obtaining a copy // of this software and associated documentation files (the "Software"), to deal // in the Software without restriction, including without limitation the rights // to use, copy, modify, merge, publish, distribute, sublicense, and/or sell // copies of the Software, and to permit persons to whom the Software is // furnished to do so, subject to the following conditions: // // The above copyright notice and this permission notice shall be included in // all copies or substantial portions of the Software. // // THE SOFTWARE IS PROVIDED "AS IS", WITHOUT WARRANTY OF ANY KIND, EXPRESS OR // IMPLIED, INCLUDING BUT NOT LIMITED TO THE WARRANTIES OF MERCHANTABILITY, // FITNESS FOR A PARTICULAR PURPOSE AND NONINFRINGEMENT. IN NO EVENT SHALL THE // AUTHORS OR COPYRIGHT HOLDERS BE LIABLE FOR ANY CLAIM, DAMAGES OR OTHER // LIABILITY, WHETHER IN AN ACTION OF CONTRACT, TORT OR OTHERWISE, ARISING FROM, // OUT OF OR IN CONNECTION WITH THE SOFTWARE OR THE USE OR OTHER DEALINGS IN // THE SOFTWARE. package cmd import ( "fmt" "github.com/spf13/cobra" ) func init() { rootCmd.AddCommand(pingCmd) rootCmd.AddCommand(listCmd) rootCmd.AddCommand(addCmd) rootCmd.AddCommand(removeCmd) removeCmd.Flags().BoolP("force", "f", false, "Forces deletion of all redirects for a hostname") } var pingCmd = &cobra.Command{ Use: "ping", Aliases: []string{"test"}, Short: "ping the server to test connection", Args: cobra.NoArgs, Hidden: true, RunE: func(cmd *cobra.Command, args []string) error { return requestFromServer("ping", args) }, } var listCmd = &cobra.Command{ Use: "list [hostname] [url]", Aliases: []string{"show"}, Short: "list the redirects on the server", Long: `list shows the redirects on a server. The command allows three forms list Lists all redirects list hostname Lists all redirects for a specific hostname list hostname url Shows the redirect for a specific hostname and url `, Example: "list www.example.com /", Args: cobra.MaximumNArgs(2), RunE: func(cmd *cobra.Command, args []string) error { return requestFromServer("list", args) }, } var addCmd = &cobra.Command{ Use: "add hostname url target", Aliases: []string{"insert", "change"}, Short: "adds a new or changes an existing redirect", Example: "add www.example.com / http://www.google.com", Args: cobra.ExactArgs(3), RunE: func(cmd *cobra.Command, args []string) error { return requestFromServer("add", args) }, } var removeCmd = &cobra.Command{ Use: "remove hostname [url]", Aliases: []string{"delete"}, Short: "remove a redirect from the server", Long: `The command allows two forms remove hostname Removes all redirects for a hostname remove hostname url Removes the redirect for a specific hostname and url `, Example: "remove www.example.com /", Args: cobra.RangeArgs(1, 2), RunE: func(cmd *cobra.Command, args []string) error { if len(args) == 1 { // full host should be removed hostname := args[0] forced, err := cmd.Flags().GetBool("force") if err != nil { return err } if !forced { fmt.Printf("Confirm to delete all redirects from hostname %v (y/n) ", hostname) var answer string _, err = fmt.Scanf("%s.1", &answer) if err != nil { return fmt.Errorf("Could not read from keyboard") } if answer != "y" && answer != "yes" { fmt.Printf("Deletion aborted for hostname %v \n", hostname) return nil } } return requestFromServer("deleteHost", args) } return requestFromServer("delete", args) }, }
<filename>calc_server/code/calc_server_add.go package main import ( "calc_util" "flag" "fmt" "log" "net/http" "strconv" ) func handler(w http.ResponseWriter, r *http.Request, url string) { output := calc_util.ResultMsg{} a, err := parseFloatQueryParamAndTransform(r, "a", url) if err != nil { output.Error = err.Error() log.Println(output.Error) } b, err := parseFloatQueryParamAndTransform(r, "b", url) if err != nil { output.Error = err.Error() log.Println(output.Error) } if output.Error == "" { output.Result = a + b } calc_util.SendOutput(w, output) } func parseFloatQueryParamAndTransform(r *http.Request, paramName, url string) (num float64, err error) { num, err = calc_util.ParseFloatQueryParam(r, paramName) if err != nil { return 0, err } num, err = calc_util.CallUrlAndReturnFloat(fmt.Sprintf("%s?a=%f", url, num)) if err != nil { return 0, err } return num, nil } func main() { port := flag.Int("port", 0, "TCP port for the HTTP server to listen on") url := flag.String("squareServerUrl", "", "URL to the 'calc_server_square' instance") flag.Parse() http.HandleFunc("/compute/add", func(w http.ResponseWriter, r *http.Request) { handler(w, r, *url) }) http.ListenAndServe(":" + strconv.Itoa(*port), nil) }
<reponame>kennethsequeira/Hello-world const HWBulgaria = () => alert('Hello World from Bulgaria! 🇧🇬'); HWBulgaria();
// @flow import Input from '../postcss/input'; import SafeParser from './safe-parser'; export default function safeParse(css, opts) { const input = new Input(css, opts); const parser = new SafeParser(input); parser.tokenize(); parser.loop(); return parser.root; }
<reponame>infinitiessoft/skyport-api<gh_stars>0 package com.infinities.skyport.cache.service; import java.io.Serializable; import java.util.concurrent.ScheduledFuture; import javax.annotation.Nullable; import com.infinities.skyport.async.service.AsyncNetworkServices; import com.infinities.skyport.async.service.network.AsyncDNSSupport; import com.infinities.skyport.async.service.network.AsyncFirewallSupport; import com.infinities.skyport.async.service.network.AsyncIpAddressSupport; import com.infinities.skyport.async.service.network.AsyncNetworkFirewallSupport; import com.infinities.skyport.async.service.network.AsyncVpnSupport; import com.infinities.skyport.cache.service.network.CachedLoadBalancerSupport; import com.infinities.skyport.cache.service.network.CachedVLANSupport; public interface CachedNetworkServices extends AsyncNetworkServices{ public static enum NetworkQuartzType implements Serializable { VLAN, POOL, SUBNET, LOADBALANCER, DNSRECORD, DNSZONE; } public @Nullable AsyncDNSSupport getDnsSupport(); public @Nullable AsyncFirewallSupport getFirewallSupport(); public @Nullable AsyncIpAddressSupport getIpAddressSupport(); public @Nullable CachedLoadBalancerSupport getLoadBalancerSupport(); public @Nullable AsyncNetworkFirewallSupport getNetworkFirewallSupport(); @Override public @Nullable CachedVLANSupport getVlanSupport(); public @Nullable AsyncVpnSupport getVpnSupport(); public ScheduledFuture<?> flushCache(NetworkQuartzType type); }
swig -python -c++ maxpooling2d.i c++ -c -fpic ../../code_test/maxpooling2d.cpp c++ -c -fpic maxpooling2d_wrap.cxx -I/usr/local/Cellar/python@3.8/3.8.5/Frameworks/Python.framework/Versions/3.8/include/python3.8 c++ -bundle -flat_namespace maxpooling2d.o maxpooling2d_wrap.o -undefined suppress -o _maxpooling2d.so
import Credential from './../models/credential'; import pageController from './page-controller'; import userTypeHelper from './../helper/credential/user-type'; import changeOnChatHelper from './../helper/employee/change-on-chat'; async function updateApiPassword(req, res) { let credential = new Credential(); credential.login = req.params['login']; credential.apiPassword = <PASSWORD>; try { let result = await credential.updateApiPassword(); return res.status(200).json({ message: result }) } catch (err) { return res.status(500).json({ message: err.message }) } } async function validateLogin(req, res) { let credential = new Credential(); credential.login = req.body.login; credential.apiPassword = <PASSWORD>; try { let credentialAuth = await credential.authenticate(); let user = await userTypeHelper.userType(credentialAuth['id']); let pages = await pageController.authorizePages(credentialAuth['id']); changeOnChatHelper.changeOnChat(credentialAuth['id'], 'on'); return res.status(200).json({ data: { "id": user['id'], "idCredential": user['id_credential'], "idRole": user['id_role'], "token": credentialAuth['token'] }, pages }) } catch (err) { return res.status(500).json({ message: err.message || 'Failed to authenticate' }) } } module.exports = { updateApiPassword: updateApiPassword, validateLogin: validateLogin }
# Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. gcloud components install cloud-datastore-emulator --quiet python3 -m venv .env . .env/bin/activate pip3 install --disable-pip-version-check -q -r requirements.txt python3 -m unittest discover -v -p "*_test.py" deactivate
<reponame>nightskylark/DevExtreme 'use strict'; var jQuery = require("jquery"); var ajax = require("../../core/utils/ajax"); var useJQuery = require("./use_jquery")(); if(useJQuery) { ajax.inject({ sendRequest: function(options) { if(!options.responseType && !options.upload) { return jQuery.ajax(options); } return this.callBase.apply(this, [options]); } }); }
function processPageRules(page, oldPage) { if (page.path !== oldPage.path) { // Assuming deletePage and createPage functions perform the respective actions deletePage(oldPage); createPage(page); return "Deleted old page, Created new page"; } if (page.path === '/senryu/show/') { page.matchPath = '/senryu/:id/'; createPage(page); return "Created new page"; } if (page.path === '/users/show/') { page.matchPath = '/users/:id/'; createPage(page); return "Created new page"; } return "No action taken"; }
package com.qht.dto; import java.util.List; import com.alibaba.fastjson.JSON; import com.alibaba.fastjson.annotation.JSONField; public class MsgBodyParameter { String GroupId; List<String> ToMembers_Account; String Content; @JSONField(name = "GroupId") public String getGroupId() { return GroupId; } public void setGroupId(String groupId) { GroupId = groupId; } @JSONField(name = "ToMembers_Account") public List<String> getToMembers_Account() { return ToMembers_Account; } public void setToMembers_Account(List<String> toMembers_Account) { ToMembers_Account = toMembers_Account; } @JSONField(name = "Content") public String getContent() { return Content; } public void setContent(String content) { Content = content; } public String toJson() { return JSON.toJSONString(this); } }
hashmap = {} hashmap['Item1'] = 'This is item 1' hashmap['Item2'] = 'This is item 2' hashmap['Item3'] = 'This is item 3' hashmap['Item4'] = 'This is item 4' hashmap['Item5'] = 'This is item 5'
#!/bin/bash echo "Starting release and build" CURRDIR="$(pwd)" export GIT_MERGE_AUTOEDIT=no ONDEVELOP="$(git branch | grep '* develop')" if [ -z "${ONDEVELOP}" ]; then echo "Must be on develop branch to get started" exit 1 fi GITSTATUS="$(git status --porcelain=1)" if [ ! -z "${GITSTATUS}" ]; then echo "No files must be uncommitted" git status exit 1 fi git pull cl_not_ok=$(grep -Pzl '## Unreleased\n+##' CHANGELOG.md) if [ "${cl_not_ok}" ]; then echo "CHANGELOG.md does not have anything entered for the Unreleased section. Edit and retry." exit 2 fi echo "Clean up assets before we start" FPHS_LOAD_APP_TYPES=1 bundle exec rake assets:clobber git commit public/assets -m "Cleanup" git push GENVERFILE=shared/build_version.txt CURRVERFILE=version.txt ALLTAGS="$(git tag --sort=-taggerdate)" CURRVER=$(cat ${CURRVERFILE}) NEWVER="$(VERSION_FILE=${CURRVERFILE} app-scripts/upversion.rb -p)" RELEASESTARTED="$(echo ${ALLTAGS} | grep ${NEWVER})" echo "Current version: ${CURRVER}" echo "Next version: ${NEWVER}" if [ -z "${SKIP_BRAKEMAN}" ]; then echo "Checking brakeman before we go through the whole process" bin/brakeman -q --summary > /tmp/fphs-brakeman-summary.txt if [ "$?" == 0 ]; then echo "Brakeman OK" else cat /tmp/fphs-brakeman-summary.txt echo "Brakeman Failed" exit 1 fi fi if [ -z "${RELEASESTARTED}" ]; then echo "Starting git-flow release" git flow release start ${NEWVER} RES=$? if [ "$RES" != "0" ]; then echo $RES exit fi git push --set-upstream origin release/${NEWVER} git flow release finish -m 'Release' ${NEWVER} else echo "Release already started. Checking out and continuing" git checkout new-master && git pull && git merge develop fi git push origin --tags git push origin --all git checkout develop echo "Starting build container" cd ../restructure-build ./build.sh if [ $? != 0 ]; then echo "***** build.sh failed with exit code $? *****" exit 101 fi if [ ! -s ${GENVERFILE} ]; then echo "${GENVERFILE} in $(pwd) was not set. The build was not successful" exit 1 fi TESTVER=$(cat ${GENVERFILE}) if [ "${TESTVER}" == "${CURRVER}" ] || [ ! "${TESTVER}" ]; then echo "Build failed" echo "'${TESTVER}' == '${CURRVER}'" exit 1 else echo "Build successful" fi cd ${CURRDIR} git fetch origin git checkout new-master git pull git checkout develop git pull git merge new-master git push echo "Built and setup assets: ${TESTVER}"
#include<iostream> #include<string.h> // Converting strings to lowercase char* toLowerCase(char* s) { int length = strlen(s); for (int i = 0; i < length; i++) { s[i] = tolower(s[i]); } return s; }
class TaskManager: def __init__(self): self.tasks = [] def delete(self, task): if not task: return -1 for index, elem in enumerate(self.tasks): if elem['task_id'] == task['task_id']: del self.tasks[index] return task['task_id'] return -1 def get(self, task): for elem in self.tasks: if elem['task_id'] == task['task_id']: return elem return None
#!/bin/sh tmux new-session -d -s 'DEV' tmux new-window -t $'DEV':1 tmux send-keys 'htop' C-m tmux split-window -h tmux send-keys 'source venv/bin/activate' C-m 'jupyter-lab --port=8001' C-m tmux split-window -v tmux send-keys 'nvidia-smi -l 1' C-m tmux -2 attach-session -d
<gh_stars>0 # Option 1 : Take the bus to the college. # Distance from the college is 5 miles. college_dist = 5 # Bus speed is 25 mph. bus_speed = 25 # Each stop delays the bus by 2 mins. stop_delay = 2 # Number of stops is 10. stops = 10 # Calculate the commute time in minutes when taking the bus. # We convert mph to mile per minute so that total_time is measured in minutes instead of hours. total_time_bus = college_dist / (bus_speed / 60) + stop_delay * stops # Print the bus commute time. print("The total commute time by bus is:", int(total_time_bus), "minutes") # Option 2 : Jog to the college. light_jog_speed = 6 heavy_jog_speed = 10 medium_jog_speed = 7 # Calculate the commute time in minutes when jogging. # We convert mph to mile per minute so that total_time is measured in minutes instead of hours. total_time_jog = 1 / (light_jog_speed / 60) + 2 / (heavy_jog_speed / 60) + 1 / (medium_jog_speed / 60) # Print the jog commute time. print("The total commute time by jogging is:", int(total_time_jog), "minutes")
#!/bin/sh chown sampledb:sampledb "${SAMPLEDB_FILE_STORAGE_PATH}" exec su sampledb -c 'env/bin/python -m sampledb "$0" "$@"' -- "$@"
<filename>loop_functions/mpga_loop_functions/mpga.cpp<gh_stars>10-100 #include "mpga.h" #include <cstdio> #include <sys/types.h> #include <sys/wait.h> #include <unistd.h> #include <sys/mman.h> #include <fcntl.h> #include <signal.h> #include <iostream> #include <fstream> #include <argos3/core/simulator/simulator.h> #include "mpga_loop_functions.h" /****************************************/ /****************************************/ /* File name for shared memory area */ static const std::string SHARED_MEMORY_FILE = "/MPGA_SHARED_MEMORY_" + ToString(getpid()); /****************************************/ /****************************************/ bool SortHighToLow(const CMPGA::SIndividual* pc_a, const CMPGA::SIndividual* pc_b) { return pc_a->Score > pc_b->Score; } bool SortLowToHigh(const CMPGA::SIndividual* pc_a, const CMPGA::SIndividual* pc_b) { return pc_b->Score > pc_a->Score; } /****************************************/ /****************************************/ CMPGA::CMPGA(const CRange<Real>& c_allele_range, UInt32 un_genome_size, UInt32 un_pop_size, Real f_mutation_prob, UInt32 un_num_trials, UInt32 un_generations, bool b_maximize, const std::string& str_argosconf, TScoreAggregator t_score_aggregator, UInt32 un_random_seed) : m_unCurrentGeneration(0), m_cAlleleRange(c_allele_range), m_unGenomeSize(un_genome_size), m_unPopSize(un_pop_size), m_fMutationProb(f_mutation_prob), m_unNumTrials(un_num_trials), m_unGenerations(un_generations), m_strARGoSConf(str_argosconf), m_tScoreAggregator(t_score_aggregator), MasterPID(::getpid()), m_cIndComparator(b_maximize ? SortHighToLow : SortLowToHigh) { /* Create shared memory manager */ m_pcSharedMem = new CSharedMem(un_genome_size, un_pop_size); /* Create slave processes */ for(UInt32 i = 0; i < m_unPopSize; ++i) { /* Perform fork */ SlavePIDs.push_back(::fork()); if(SlavePIDs.back() == 0) { /* We're in a slave */ LaunchARGoS(i); } } /* Create a random number generator */ CRandom::CreateCategory("ga", un_random_seed); m_pcRNG = CRandom::CreateRNG("ga"); /* Create initial population */ SIndividual* psInd; for(size_t p = 0; p < m_unPopSize; ++p) { /* Create individual */ psInd = new SIndividual; psInd->Score = -1.0; /* Create random genome */ for(size_t g = 0; g < m_unGenomeSize; ++g) { psInd->Genome.push_back(m_pcRNG->Uniform(m_cAlleleRange)); } /* Add individual to the population */ m_tPopulation.push_back(psInd); } /* The master sleeps to give enough time to the slaves to * initialize and suspend properly. If not enough time is given * here, the master will hang later on. */ ::sleep(3); } /****************************************/ /****************************************/ CMPGA::~CMPGA() { /* Terminate slaves */ for(UInt32 i = 0; i < m_unPopSize; ++i) { ::kill(SlavePIDs[i], SIGTERM); } /* Clean memory up */ while(!m_tPopulation.empty()) { delete m_tPopulation.back(); m_tPopulation.pop_back(); } CRandom::RemoveCategory("ga"); /* Other cleanup in common between master and slaves */ Cleanup(); } /****************************************/ /****************************************/ const CMPGA::TPopulation& CMPGA::GetPopulation() const { return m_tPopulation; } /****************************************/ /****************************************/ UInt32 CMPGA::GetGeneration() const { return m_unCurrentGeneration; } /****************************************/ /****************************************/ void CMPGA::Cleanup() { delete m_pcSharedMem; } /****************************************/ /****************************************/ void CMPGA::Evaluate() { /* Set parameters for the processes and resume them */ for(UInt32 i = 0; i < m_unPopSize; ++i) { /* Set genome */ m_pcSharedMem->SetGenome(i, &(m_tPopulation[i]->Genome[0])); /* Resume process */ ::kill(SlavePIDs[i], SIGCONT); } /* Wait for all the slaves to finish the run */ UInt32 unTrialsLeft = m_unPopSize; int nSlaveInfo; pid_t tSlavePID; while(unTrialsLeft > 0) { /* Wait for next slave to finish */ tSlavePID = ::waitpid(-1, &nSlaveInfo, WUNTRACED); /* Make sure the slave went back to sleep and didn't crash */ if(!WIFSTOPPED(nSlaveInfo)) { LOGERR << "[FATAL] Slave process with PID " << tSlavePID << " exited, can't continue. Check file ARGoS_LOGERR_" << tSlavePID << " for more information." << std::endl; LOG.Flush(); LOGERR.Flush(); Cleanup(); ::exit(1); } /* All OK, one less slave to wait for */ --unTrialsLeft; } /* Copy the scores into the population data */ for(UInt32 i = 0; i < m_unPopSize; ++i) { m_tPopulation[i]->Score = m_pcSharedMem->GetScore(i); } /* Sort the population by score, from the best to the worst */ std::sort(m_tPopulation.begin(), m_tPopulation.end(), m_cIndComparator); } /****************************************/ /****************************************/ void CMPGA::NextGen() { ++m_unCurrentGeneration; Selection(); Crossover(); Mutation(); } /****************************************/ /****************************************/ bool CMPGA::Done() const { return m_unCurrentGeneration >= m_unGenerations; } /****************************************/ /****************************************/ /* Global pointer to the CMPGA object in the current slave, used by * SlaveHandleSIGTERM() to perform cleanup */ static CMPGA* GA_INSTANCE; /* SIGTERM handler for slave processes */ void SlaveHandleSIGTERM(int) { argos::CSimulator::GetInstance().Destroy(); argos::LOG.Flush(); argos::LOGERR.Flush(); GA_INSTANCE->Cleanup(); } void CMPGA::LaunchARGoS(UInt32 un_slave_id) { /* Set the global GA instance pointer for signal handler */ GA_INSTANCE = this; /* Install handler for SIGTERM */ ::signal(SIGTERM, SlaveHandleSIGTERM); /* Initialize ARGoS */ /* Redirect LOG and LOGERR to dedicated files to prevent clutter on the screen */ std::ofstream cLOGFile(std::string("ARGoS_LOG_" + ToString(::getpid())).c_str(), std::ios::out); LOG.DisableColoredOutput(); LOG.GetStream().rdbuf(cLOGFile.rdbuf()); std::ofstream cLOGERRFile(std::string("ARGoS_LOGERR_" + ToString(::getpid())).c_str(), std::ios::out); LOGERR.DisableColoredOutput(); LOGERR.GetStream().rdbuf(cLOGERRFile.rdbuf()); /* The CSimulator class of ARGoS is a singleton. Therefore, to * manipulate an ARGoS experiment, it is enough to get its instance */ argos::CSimulator& cSimulator = argos::CSimulator::GetInstance(); try { /* Set the .argos configuration file * This is a relative path which assumed that you launch the executable * from argos3-examples (as said also in the README) */ cSimulator.SetExperimentFileName(m_strARGoSConf); /* Load it to configure ARGoS */ cSimulator.LoadExperiment(); LOG.Flush(); LOGERR.Flush(); } catch(CARGoSException& ex) { LOGERR << ex.what() << std::endl; ::raise(SIGTERM); } /* Get a reference to the loop functions */ CMPGALoopFunctions& cLoopFunctions = dynamic_cast<CMPGALoopFunctions&>(cSimulator.GetLoopFunctions()); /* Create vector of scores */ std::vector<Real> vecScores(m_unNumTrials, 0.0); /* Continue working until killed by parent */ while(1) { /* Suspend yourself, waiting for parent's resume signal */ ::raise(SIGTSTP); /* Resumed */ /* Configure the controller with the genome */ cLoopFunctions.ConfigureFromGenome(m_pcSharedMem->GetGenome(un_slave_id)); /* Run the trials */ for(size_t i = 0; i < m_unNumTrials; ++i) { /* Tell the loop functions to get ready for the i-th trial */ cLoopFunctions.SetTrial(i); /* Reset the experiment. * This internally calls also CMPGALoopFunctions::Reset(). */ cSimulator.Reset(); /* Run the experiment */ cSimulator.Execute(); /* Store score */ vecScores[i] = cLoopFunctions.Score(); LOG.Flush(); LOGERR.Flush(); } ; /* Put result in shared memory */ m_pcSharedMem->SetScore(un_slave_id, m_tScoreAggregator(vecScores)); } } /****************************************/ /****************************************/ void CMPGA::Selection() { /* Delete all individuals apart from the top two */ while(m_tPopulation.size() > 2) { delete m_tPopulation.back(); m_tPopulation.pop_back(); } } /****************************************/ /****************************************/ void CMPGA::Crossover() { /* * This is a simple one-point crossover. */ SIndividual* psParent1 = m_tPopulation[0]; SIndividual* psParent2 = m_tPopulation[1]; UInt32 unCut; SIndividual* psInd; for(UInt32 i = 2; i < m_unPopSize; ++i) { /* Pick a cutting point at random */ unCut = m_pcRNG->Uniform(CRange<UInt32>(1, m_unGenomeSize-1)); /* Make a new individual */ psInd = new SIndividual; /* Copy alleles from parent 1 */ for(UInt32 j = 0; j < unCut; ++j) { psInd->Genome.push_back(psParent1->Genome[j]); } /* Copy alleles from parent 2 */ for(UInt32 j = unCut; j < m_unGenomeSize; ++j) { psInd->Genome.push_back(psParent2->Genome[j]); } /* Add individual to the new population */ m_tPopulation.push_back(psInd); } } /****************************************/ /****************************************/ void CMPGA::Mutation() { /* Mutate the alleles of the newly added individuals by setting a * new random value from a uniform distribution */ for(UInt32 i = 2; i < m_unPopSize; ++i) { for(UInt32 a = 0; a < m_unGenomeSize; ++a) { if(m_pcRNG->Bernoulli(m_fMutationProb)) m_tPopulation[i]->Genome[a] = m_pcRNG->Uniform(m_cAlleleRange); } } } /****************************************/ /****************************************/ CMPGA::CSharedMem::CSharedMem(UInt32 un_genome_size, UInt32 un_pop_size) : m_unGenomeSize(un_genome_size), m_unPopSize(un_pop_size) { /* Create shared memory area for master-slave communication */ m_nSharedMemFD = ::shm_open(SHARED_MEMORY_FILE.c_str(), O_RDWR | O_CREAT, S_IRUSR | S_IWUSR); if(m_nSharedMemFD < 0) { ::perror(SHARED_MEMORY_FILE.c_str()); exit(1); } /* Resize shared memory area to contain the population data * - The area must contain m_unPopSize elements * - Each element must have space for the data of an individual * - Genome: m_unGenomeSize * sizeof(Real) * - Score: sizeof(Real) */ size_t unShareMemSize = m_unPopSize * (m_unGenomeSize+1) * sizeof(Real); ::ftruncate(m_nSharedMemFD, unShareMemSize); /* Get pointer to shared memory area */ m_pfSharedMem = reinterpret_cast<Real*>( ::mmap(NULL, unShareMemSize, PROT_READ | PROT_WRITE, MAP_SHARED, m_nSharedMemFD, 0)); if(m_pfSharedMem == MAP_FAILED) { ::perror("shared memory"); exit(1); } } /****************************************/ /****************************************/ CMPGA::CSharedMem::~CSharedMem() { munmap(m_pfSharedMem, m_unPopSize * (m_unGenomeSize+1) * sizeof(Real)); close(m_nSharedMemFD); shm_unlink(SHARED_MEMORY_FILE.c_str()); } /****************************************/ /****************************************/ Real* CMPGA::CSharedMem::GetGenome(UInt32 un_individual) { return m_pfSharedMem + un_individual * (m_unGenomeSize+1); } /****************************************/ /****************************************/ void CMPGA::CSharedMem::SetGenome(UInt32 un_individual, const Real* pf_genome) { ::memcpy(m_pfSharedMem + un_individual * (m_unGenomeSize+1), pf_genome, m_unGenomeSize * sizeof(Real)); } /****************************************/ /****************************************/ Real CMPGA::CSharedMem::GetScore(UInt32 un_individual) { return m_pfSharedMem[un_individual * (m_unGenomeSize+1) + m_unGenomeSize]; } /****************************************/ /****************************************/ void CMPGA::CSharedMem::SetScore(UInt32 un_individual, Real f_score) { m_pfSharedMem[un_individual * (m_unGenomeSize+1) + m_unGenomeSize] = f_score; } /****************************************/ /****************************************/
#!/usr/bin/env bash # # Copyright (c) 2021 SAP SE or an SAP affiliate company. All rights reserved. This file is licensed under the Apache Software License, v. 2 except as noted otherwise in the LICENSE file # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. set -o errexit set -o nounset set -o pipefail LOCAL_GARDEN_LABEL=${1:-local-garden} ACTIVATE_SEEDAUTHORIZER=${2:-false} KUBECONFIGPATH="$(dirname $0)/kubeconfigs/default-admin.conf" echo "# Remove old containers and create the docker user network" $(dirname $0)/cleanup $LOCAL_GARDEN_LABEL docker network create gardener-dev --label $LOCAL_GARDEN_LABEL echo "# Start the nodeless kubernetes environment" $(dirname $0)/run-kube-etcd $LOCAL_GARDEN_LABEL $(dirname $0)/run-kube-apiserver $LOCAL_GARDEN_LABEL $ACTIVATE_SEEDAUTHORIZER $(dirname $0)/run-kube-controller-manager $LOCAL_GARDEN_LABEL echo "# This etcd will be used to store gardener resources (e.g., seeds, shoots)" $(dirname $0)/run-gardener-etcd $LOCAL_GARDEN_LABEL for i in 1..10; do if $(KUBECONFIG=$KUBECONFIGPATH kubectl cluster-info > /dev/null 2>&1); then break fi echo "# Waiting until Kube-Apiserver is available" done echo "# Configuring RBAC resources for Gardener components" $(dirname $0)/configure-rbac echo "# Applying proxy RBAC for the extension controller" echo "# After this step, you can start using the cluster at KUBECONFIG=hack/local-development/local-garden/kubeconfigs/default-admin.conf" $(dirname $0)/apply-rbac-garden-ns echo "# Now you can start using the cluster at with \`export KUBECONFIG=hack/local-development/local-garden/kubeconfigs/default-admin.conf\`" echo "# Then you need to run \`make dev-setup\` to setup config and certificates files for gardener's components and to register the gardener-apiserver." echo "# Finally, run \`make start-apiserver,start-controller-manager,start-scheduler,start-gardenlet\` to start the gardener components as usual." if $ACTIVATE_SEEDAUTHORIZER; then echo '# Additionally, make sure to run make start-admission-controller' fi
#!/usr/bin/env bash # Copyright 2018 The Kubernetes Authors. # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # simple script to build binaries for release set -o errexit -o nounset -o pipefail # cd to the repo root REPO_ROOT="$(cd "$(dirname "${BASH_SOURCE[0]}")/../../.." && pwd -P)" cd "${REPO_ROOT}" # controls the number of concurrent builds PARALLELISM=${PARALLELISM:-6} # trivial go_container.sh command to ensure the volumes are setup # before we start running many containers in parallel # https://github.com/kubernetes-sigs/kind/issues/974 hack/go_container.sh go version echo "Building in parallel for:" # What we do here: # - use xargs to build in parallel (-P) while collecting a combined exit code # - use cat to supply the individual args to xargs (one line each) # - use env -S to split the line into environment variables and execute # - ... the build # NOTE: the binary name needs to be in single quotes so we delay evaluating # GOOS / GOARCH # NOTE: disable SC2016 because we _intend_ for these to evaluate later # shellcheck disable=SC2016 if xargs -0 -n1 -P "${PARALLELISM}" bash -c 'eval $0; make build KIND_BINARY_NAME=kind-${GOOS}-${GOARCH}'; then echo "Cross build passed!" 1>&2 else echo "Cross build failed!" 1>&2 exit 1 fi < <(cat <<EOF | tr '\n' '\0' export GOOS=windows GOARCH=amd64 export GOOS=darwin GOARCH=amd64 export GOOS=linux GOARCH=amd64 export GOOS=linux GOARCH=arm64 export GOOS=linux GOARCH=ppc64le EOF )
'use strict'; const { getResolvedSchema, getParentShapeInForm, getValue, printSchema, handleResponse } = require('@bbp/nexus-shacl-helpers'); const createForm = require('./form-generation'); const { fillForm, sendForm } = require('./form-actions'); const config = require('../config'); let rev; const urlParts = location.pathname.split('/'); urlParts.splice(0, 3); if (urlParts.length === 5) { const ID = urlParts[4]; const ver = urlParts[3]; const schema = urlParts[2]; const domain = urlParts[1]; const org = urlParts[0]; const dataUrl = `${config.endpoints.dataPrefix }/${org}/${domain}/${schema}/${ver}/${ID}`; const schemaUrl = `${config.endpoints.schemasPrefix}/${org}/${domain}/${schema}/${ver}`; fetch(dataUrl) .then(handleResponse) .then(data => { rev = data.rev; fetch(schemaUrl) .then(handleResponse) .then(getResolvedSchema) .then(schema => { return printSchema({content: schema, label: 'Schema details'}); }) .then(createForm) .then(schema => { fillForm(schema, data); return data; }) .then(data => { return printSchema({content: data, label: 'Instance details'}); }) }) .catch(err => { alert('This Instance doesnt exist'); console.log(err); }) } else { alert('specify the ID'); } document.body.addEventListener('click', function(event) { if (event.target.type === 'submit') { event.preventDefault(); sendForm(ID, rev); } }, false);
#!/usr/bin/env bash ## ************************************************************************* # Deployment script for Magento 2 based apps. # # This is friendly user script, not user friendly # There are no protection from mistakes. # Use it if you know how it works. ## ************************************************************************* # pin current folder and deployment root folder DIR_CUR="$PWD" DIR_ROOT="$( cd "$( dirname "$0" )" && pwd )" # default configuration OPT_CLI_HELP="no" # -h print out help OPT_SKIP_DB="no" # -S OPT_USE_EXIST_DB="no" # -E OPT_MAGE_RUN=developer # -m developer|production OPT_MODE=work # -d work|live # Available deployment modes MODE_LIVE=live MODE_WORK=work ## ************************************************************************* # Parse input options ## ************************************************************************* echo "" while getopts "d:hm:DEMS" OPTNAME do case "${OPTNAME}" in "d") OPT_MODE=${OPTARG} echo "Application deployment mode '${OPT_MODE}' is specified." ;; "h") OPT_CLI_HELP="yes" ;; "m") OPT_MAGE_RUN=${OPTARG} echo "Magento deployment mode '${OPT_MAGE_RUN}' is specified." ;; "E") OPT_USE_EXIST_DB="yes" echo "Existing DB will be used in 'work' mode." ;; "S") OPT_SKIP_DB="yes" echo "Database initialization will be skipped (use Web UI to create DB)." ;; esac done echo"" ## ************************************************************************* # Print out help ## ************************************************************************* if [ "${OPT_CLI_HELP}" = "yes" ]; then echo "Magento2 application deployment script." echo "" echo "Usage: sh deploy.sh -d [work|live] -h -m [developer|production] -E -S" echo "" echo "Where:" echo " -d: Web application deployment mode ([work|live], default: work);" echo " -h: This output;" echo " -m: Magento 2 itself deployment mode ([developer|production], default: developer);" echo " -E: Existing DB will be used;" echo " -S: Skip database initialization (Web UI should be used to init DB);" exit fi ## ************************************************************************* # Validate current deployment mode (work|live) ## ************************************************************************* MODE=${MODE_WORK} case "${OPT_MODE}" in ${MODE_WORK}|${MODE_LIVE}) MODE=${OPT_MODE};; esac # check configuration file exists and load deployment config (db connection, Magento installation opts, etc.). FILE_CFG=${DIR_ROOT}/cfg.${MODE}.sh if [ -f "${FILE_CFG}" ] then echo "There is deployment configuration in ${FILE_CFG}." . ${FILE_CFG} else echo "There is no expected configuration in ${FILE_CFG}. Aborting..." cd ${DIR_CUR} exit 255 fi # shortcuts to external vars (from ./cfg.${MODE}.sh) BASE_URL=${BASE_URL} BACKEND_FRONTNAME=${BACKEND_FRONTNAME} ## ************************************************************************* # Refresh deployment script ## ************************************************************************* git pull ## ************************************************************************* # Deployment process itself ## ************************************************************************* cd ${DIR_ROOT} . ./deploy/bin/app.sh echo "" cd ${DIR_ROOT} . ./deploy/bin/final.sh echo "" echo "Application deployment in '${OPT_MODE}' mode is done." echo " Frontend: ${BASE_URL}" echo " Backend: ${BASE_URL}/${BACKEND_FRONTNAME}" echo "" cd ${DIR_CUR}
import json def write_section_annos_to_json(section_annos, json_file): with open(json_file, 'w') as outfile: json.dump(section_annos, outfile, indent=4)
<reponame>davidleiva/folio_portfolio import React, { useContext } from 'react' import styled from 'styled-components' import { Container, Row, Col } from 'react-bootstrap' import GlobalContext from '../../context/GlobalContext' import { Section, Title, ButtonIcon } from '../../components/Core' import Availability from '../../components/Availability' import bgFooter from '../../assets/image/png/subscribe-pattern.png' const SectionStyled = styled(Section)` &::before { position: absolute; bottom: -150px; content: ''; width: 120%; height: 150%; background: url(${bgFooter}) bottom center no-repeat; background-size: cover; z-index: -1; } ` const Hero = () => { const gContext = useContext(GlobalContext) return ( <> <SectionStyled className="position-relative"> <Container> <Row className="text-center justify-content-center"> <Col lg="11" xl="8"> <Availability /> <div className="text-center my-5"> <Title>Give your product a shape</Title> </div> <div className="text-center"> <ButtonIcon onClick={(e) => { e.preventDefault() gContext.toggleContact() }} > Let's Talk Now </ButtonIcon> </div> </Col> </Row> </Container> </SectionStyled> </> ) } export default Hero
package com.sbsuen.fitfam.user; import org.springframework.data.mongodb.repository.MongoRepository; public interface UserRepository extends MongoRepository<User,String> { }
#!/bin/bash # trova tutti gli enigmi di un certo autore # se un autore contiene spazi deve essere racchiuso tra virgolette AUTORE=$(echo $1 | sed -e "s/ /%20/g") echo "# tutti gli enigmi dell'autore $1" echo $(curl -s localhost:8080/enigmi/cercaenigmi/autore/$AUTORE) echo
<?php namespace Property; class Location { // Implementation of Property\Location class } class MainEntityOfPage { // Implementation of Property\MainEntityOfPage class } class Name { // Implementation of Property\Name class } class ReceiveAction { private $location; private $mainEntityOfPage; private $name; public function setLocation(Property\Location $location) { $this->location = $location; } public function setMainEntityOfPage(Property\MainEntityOfPage $mainEntityOfPage) { $this->mainEntityOfPage = $mainEntityOfPage; } public function setName(Property\Name $name) { $this->name = $name; } } // Usage $receiveAction = new ReceiveAction(); $location = new Property\Location(); $mainEntityOfPage = new Property\MainEntityOfPage(); $name = new Property\Name(); $receiveAction->setLocation($location); $receiveAction->setMainEntityOfPage($mainEntityOfPage); $receiveAction->setName($name);
/* * Copyright 2011 Google Inc. * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package com.google.enterprise.secmgr.saml; import org.opensaml.common.SAMLObject; import javax.annotation.Nonnull; import javax.annotation.ParametersAreNonnullByDefault; import javax.xml.XMLConstants; import javax.xml.namespace.QName; /** * A SAML extension added to AuthnRequest messages by the GSA. */ @ParametersAreNonnullByDefault public interface GsaAuthn extends SAMLObject { public static final QName DEFAULT_ELEMENT_NAME = new QName(OpenSamlUtil.GOOGLE_NS_URI, "GsaAuthn", OpenSamlUtil.GOOGLE_NS_PREFIX); public static final QName VERSION_ATTRIB_NAME = new QName(XMLConstants.NULL_NS_URI, "version", XMLConstants.DEFAULT_NS_PREFIX); public static final QName SESSION_ID_ATTRIB_NAME = new QName(XMLConstants.NULL_NS_URI, "sessionId", XMLConstants.DEFAULT_NS_PREFIX); public static final int MIN_VERSION = 1; public static final int CURRENT_VERSION = 1; public static final int MAX_VERSION = CURRENT_VERSION; public int getVersion(); public void setVersion(int version); @Nonnull public String getSessionId(); public void setSessionId(String sessionId); }
import string import random def generate_password(min_length=8, max_length=16): # generate random password from the given options random_password = '' pool = string.ascii_letters + string.digits + string.punctuation # randomly select characters from the pool and combine them # to form the password length = random.randint(min_length, max_length) for i in range(length): random_password += random.choice(pool) # return the generated password return random_password if __name__ == '__main__': print(generate_password())
package visao; import java.io.IOException; import javafx.application.Application; import javafx.fxml.FXMLLoader; import javafx.scene.Parent; import javafx.scene.Scene; import javafx.stage.Stage; import javafx.stage.StageStyle; public class aplicacao extends Application { private Parent parent; private FXMLLoader fxmlLoader; private static FXMLDocumentController controller; @Override public void start(Stage stage) throws Exception { Parent parent = null; FXMLLoader fxmlLoader = null; try { fxmlLoader = new FXMLLoader(getClass().getResource("FXMLDocument.fxml")); parent = (Parent) fxmlLoader.load(); } catch (IOException ex) { System.out.println("Erro ao iniciar a tela de cadastro de Estado "+ex.getMessage()); } controller = (FXMLDocumentController) fxmlLoader.getController(); Scene scene = new Scene(parent); stage.setScene(scene); stage.setTitle("Lavoro"); stage.setResizable(false); stage.show(); } public static FXMLDocumentController getController(){ return controller; } public static void main(String[] args) { launch(args); } }
/* global game Phaser game_state */ game_state.end = function() {}; game_state.end.prototype = { preload: function() { }, create: function() { this.scoreText = game.add.text(16, 16, "Game Over", { fontSize: '64px', fill: '#ffffff' }); }, update: function() { }, } game.state.add('end', game_state.end); game.state.start('end');
#!/bin/bash # Copyright (c) Meta Platforms, Inc. and affiliates. # # This source code is licensed under the MIT license found in the # LICENSE file in the root directory of this source tree. assert_errors "$1" check . --all --no-flowlib --show-all-errors --include-warnings --color=always --unicode=always
<reponame>gabmontes/node-coindesk-api if (!Array.prototype.includes) { require('core-js/fn/array/includes') } const request = require('./request') const formatDate = require('./formatDate') // memoize 1' function getSupportedCurrencies() { return request('/supported-currencies.json') } // memoize 1' TTL 15" function getCurrentPrice(currency) { if (!currency) { return request('/currentprice.json') } return request(`/currentprice/${currency}.json`) } // memoize 1' TTL 15" function getHistoricalClosePrices(options = {}) { const { index, currency, start, end, yesterday } = options return request(`/historical/close.json`, { index, currency, start: start && end && formatDate(start) || undefined, end: start && end && formatDate(end) || undefined, for: yesterday ? 'yesterday' : undefined }) } module.exports = { getCurrentPrice, getSupportedCurrencies, getHistoricalClosePrices }
<filename>observatory-platform/observatory/platform/cli/platform_command.py # Copyright 2020 Curtin University # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. # Author: <NAME> from observatory.platform.observatory_config import BackendType from observatory.platform.platform_builder import PlatformBuilder, HOST_UID, DEBUG from observatory.platform.utils.url_utils import wait_for_url class PlatformCommand(PlatformBuilder): def __init__(self, config_path: str, host_uid: int = HOST_UID, debug: bool = DEBUG): """Create a PlatformCommand, which can be used to start and stop Observatory Platform instances. :param config_path: The path to the config.yaml configuration file. :param host_uid: The user id of the host system. Used to set the user id in the Docker containers. :param debug: Print debugging information. """ super().__init__( config_path=config_path, host_uid=host_uid, debug=debug, backend_type=BackendType.local ) @property def ui_url(self) -> str: """Return the URL to Apache Airflow UI. :return: Apache Airflow UI URL. """ return f"http://localhost:{self.config.observatory.airflow_ui_port}" def wait_for_airflow_ui(self, timeout: int = 60) -> bool: """Wait for the Apache Airflow UI to start. :param timeout: the number of seconds to wait before timing out. :return: whether connecting to the Apache Airflow UI was successful or not. """ return wait_for_url(self.ui_url, timeout=timeout)
def even_number_filter(number): return number % 2 == 0 numbers = [25, 35, 8, 2, 10] filtered_list = list(filter(even_number_filter, numbers)) print(filtered_list)
<gh_stars>1-10 #include <errno.h> #include <hiredis/hiredis.h> #include <poll.h> #include <pthread.h> #include <stdbool.h> #include <stdlib.h> #include <string.h> #include <unistd.h> #include "thredis.h" struct redis_wait { struct redis_wait* next; redisReply* reply; pthread_mutex_t mutex; pthread_cond_t done; }; struct thredis { redisContext* redis; pthread_mutex_t mutex; pthread_t reader_thread; struct redis_wait* wait_head; struct redis_wait** wait_tail; }; static struct redis_wait* pop_waiter(thredis_t* thredis) { struct redis_wait* wait = thredis->wait_head; thredis->wait_head = wait->next; if (thredis->wait_head == NULL) { thredis->wait_tail = &thredis->wait_head; } return wait; } static void push_waiter(thredis_t* thredis, struct redis_wait* waiter) { *thredis->wait_tail = waiter; thredis->wait_tail = &waiter->next; } static void deliver_reply(thredis_t* thredis, redisReply* reply) { struct redis_wait* wait = pop_waiter(thredis); wait->reply = reply; pthread_mutex_lock(&wait->mutex); pthread_cond_signal(&wait->done); pthread_mutex_unlock(&wait->mutex); } static void* reader_thread_main(void* ctx) { thredis_t* thredis = ctx; struct pollfd pfd = { .fd = thredis->redis->fd, .events = POLLIN }; while (1) { int rc = poll(&pfd, 1, -1); if (rc < 0 && errno == EINTR) { continue; } if (pfd.revents & POLLERR) { // TODO - how do we tell the user about this error? pthread_mutex_lock(&thredis->mutex); while (thredis->wait_head) { deliver_reply(thredis, NULL); } thredis->redis = NULL; pthread_mutex_unlock(&thredis->mutex); return NULL; } if (pfd.revents & POLLIN) { pthread_mutex_lock(&thredis->mutex); if (redisBufferRead(thredis->redis) != REDIS_OK) { // TODO - how do we tell the user about this error? while (thredis->wait_head) { deliver_reply(thredis, NULL); } thredis->redis = NULL; pthread_mutex_unlock(&thredis->mutex); return NULL; } while (1) { redisReply* reply = NULL; if (redisGetReplyFromReader(thredis->redis, (void**) &reply) != REDIS_OK) { // TODO - how do we tell the user about this error? while (thredis->wait_head) { deliver_reply(thredis, NULL); } thredis->redis = NULL; pthread_mutex_unlock(&thredis->mutex); return NULL; } if (reply == NULL) { break; } deliver_reply(thredis, reply); } pthread_mutex_unlock(&thredis->mutex); } } } static int flush_writes(redisContext* redis) { int done = 0; while (!done) { int rc = redisBufferWrite(redis, &done); if (rc != REDIS_OK) { return rc; } } return REDIS_OK; } thredis_t* thredis_new(redisContext* redis) { thredis_t* thredis = malloc(sizeof(*thredis)); if (!thredis) { return NULL; } thredis->redis = redis; thredis->wait_head = NULL; thredis->wait_tail = &thredis->wait_head; if (pthread_mutex_init(&thredis->mutex, NULL)) { free(thredis); return NULL; } if (pthread_create(&thredis->reader_thread, NULL, reader_thread_main, thredis)) { pthread_mutex_destroy(&thredis->mutex); free(thredis); return NULL; } return thredis; } void thredis_close(thredis_t* thredis) { pthread_mutex_lock(&thredis->mutex); pthread_cancel(thredis->reader_thread); pthread_join(thredis->reader_thread, NULL); pthread_mutex_unlock(&thredis->mutex); pthread_mutex_destroy(&thredis->mutex); free(thredis); } redisReply* thredis_command(thredis_t* thredis, const char* format, ...) { va_list va; va_start(va, format); pthread_mutex_lock(&thredis->mutex); if (!thredis->redis) { pthread_mutex_unlock(&thredis->mutex); va_end(va); return NULL; } if (redisvAppendCommand(thredis->redis, format, va) != REDIS_OK) { pthread_mutex_unlock(&thredis->mutex); va_end(va); return NULL; } if (flush_writes(thredis->redis) != REDIS_OK) { pthread_mutex_unlock(&thredis->mutex); va_end(va); return NULL; } struct redis_wait wait = { .next = NULL, .reply = NULL, }; pthread_mutex_init(&wait.mutex, NULL); pthread_cond_init(&wait.done, NULL); push_waiter(thredis, &wait); pthread_mutex_lock(&wait.mutex); pthread_mutex_unlock(&thredis->mutex); va_end(va); pthread_cond_wait(&wait.done, &wait.mutex); pthread_mutex_unlock(&wait.mutex); pthread_cond_destroy(&wait.done); pthread_mutex_destroy(&wait.mutex); return wait.reply; }
module.exports = { run: () => 'implied' }
echo '[+] Installing Dependencies...' pkg update pkg upgrade echo '[!]Python' pkg install python echo '[!]PHP' pkg install php echo '[!]wget' pkg install wget echo '[!]unzip' pkg install unzip echo '[!]openssh' pkg install ssh echo '[+]Requests' pip install requests echo '[+] Installed.'
<reponame>Blockception/BC-Minecraft-Bedrock-Vanilla-Data /* Auto generated */ export * from "./BehaviorPack"; export * from "./Block"; export * from "./Entity"; export * from "./Item"; export * from "./LootTable"; export * from "./Trading";
/* * Copyright 2018 <NAME> * * Licensed under the Apache License, Version 2.0 (the "License"); * you may not use this file except in compliance with the License. * You may obtain a copy of the License at * * http://www.apache.org/licenses/LICENSE-2.0 * * Unless required by applicable law or agreed to in writing, software * distributed under the License is distributed on an "AS IS" BASIS, * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. * See the License for the specific language governing permissions and * limitations under the License. */ package flix.runtime.fixpoint.predicate; import flix.runtime.fixpoint.symbol.PredSym; import flix.runtime.fixpoint.term.LitTerm; import flix.runtime.fixpoint.term.Term; import java.util.Arrays; import java.util.StringJoiner; /** * Represents an atom predicate of the form: sym(terms). */ public final class AtomPredicate implements Predicate { /** * Constructs an atom predicate for the given predicate symbol, with the given polarity, and terms. */ public static AtomPredicate of(PredSym sym, boolean positive, Term[] terms) { if (sym == null) throw new IllegalArgumentException("'sym' must be non-null."); if (terms == null) throw new IllegalArgumentException("'terms' must be non-null."); return new AtomPredicate(sym, positive, terms); } /** * The predicate symbol. */ private final PredSym sym; /** * Whether the atom is negated. */ private final boolean positive; /** * The terms of the atom. */ private final Term[] terms; /** * Private constructor. */ private AtomPredicate(PredSym sym, boolean positive, Term[] terms) { this.sym = sym; this.positive = positive; this.terms = terms; } /** * Returns the symbol of `this` atom. */ public PredSym getSym() { return sym; } /** * Returns `true` if `this` atom is positive. */ public Boolean isPositive() { return positive; } /** * Returns `true` if `this` atom is negative. */ public Boolean isNegative() { return !isPositive(); } /** * Returns the terms of `this` atom. */ public Term[] getTerms() { return terms; } /** * Returns `true` if `this` atom is ground. */ public boolean isGround() { for (Term t : terms) { if (!(t instanceof LitTerm)) { return false; } } return true; } /** * Returns `true` if `this` atom entails `that` atom. */ public boolean entails(AtomPredicate that) { if (that == null) throw new IllegalArgumentException("'that' must be non-null."); if (!this.isGround()) { throw new IllegalArgumentException("'this' must be ground."); } if (!that.isGround()) { throw new IllegalArgumentException("'that' must be ground."); } if (!this.sym.equals(that.sym)) { // Case 1: Symbols differ. return false; } // TODO: Lattice semantics. for (var i = 0; i < this.getTerms().length; i++) { var thisTerm = (LitTerm) this.getTerms()[i]; var thatTerm = (LitTerm) that.getTerms()[i]; var thisLit = thisTerm.getFunction().apply(new Object[1]); var thatLit = thatTerm.getFunction().apply(new Object[1]); if (!thisLit.equals(thatLit)) { // Case 2: A literal differs. return false; } } return true; } /** * Returns a human-readable representation of `this` predicate. */ @Override public String toString() { StringJoiner sj = new StringJoiner(",", sym.toString() + "(", ")"); for (var term : terms) { sj.add(term.toString()); } return sj.toString(); } }
#[derive(Copy, Clone, Debug)] pub enum MobTargetMode { NearbyCell = 0, NearbyPlayer = 1, Clockwise = 2, Anticlockwise = 3, ClockwiseNext = 4, } #[derive(Copy, Clone, Debug)] pub enum Direction { North, East, South, West, } pub fn calculate_next_move(current_position: (i32, i32), current_direction: Direction, target_mode: MobTargetMode) -> (i32, i32) { match target_mode { MobTargetMode::NearbyCell => { // For NearbyCell mode, the mob can move to any adjacent cell // For simplicity, let's assume the mob moves to the cell directly in front of it match current_direction { Direction::North => (current_position.0, current_position.1 - 1), Direction::East => (current_position.0 + 1, current_position.1), Direction::South => (current_position.0, current_position.1 + 1), Direction::West => (current_position.0 - 1, current_position.1), } } MobTargetMode::NearbyPlayer => { // Implement logic to follow a nearby player // This could involve tracking the player's position and moving towards it unimplemented!() } MobTargetMode::Clockwise => { // Implement logic to move in a clockwise direction // This could involve rotating the direction of movement in a clockwise manner unimplemented!() } MobTargetMode::Anticlockwise => { // Implement logic to move in an anticlockwise direction // This could involve rotating the direction of movement in an anticlockwise manner unimplemented!() } MobTargetMode::ClockwiseNext => { // Implement logic to move in a clockwise direction starting from the next direction // This could involve rotating the direction of movement in a clockwise manner, starting from the next direction unimplemented!() } } }
<filename>SOLVER/src/core/output/element-wise/eigen_element_op.hpp // // eigen_element_op.hpp // AxiSEM3D // // Created by <NAME> on 28/7/20. // Copyright © 2020 <NAME>. All rights reserved. // // eigen for element output #ifndef eigen_element_op_hpp #define eigen_element_op_hpp #include "eigen_station.hpp" #include "eigen_generic.hpp" namespace eigen { using Eigen::Dynamic; using numerical::Real; using spectral::nPEM; // tensor typedef Eigen::Tensor<numerical::Real, 5, Eigen::RowMajor> RTensor5; typedef Eigen::Tensor<numerical::Real, 4, Eigen::RowMajor> RTensor4; typedef Eigen::array<Eigen::DenseIndex, 5> IArray5; typedef Eigen::array<Eigen::DenseIndex, 4> IArray4; // element-na info typedef Eigen::Matrix<int, Eigen::Dynamic, 4, Eigen::RowMajor> IMatX4_RM; typedef Eigen::Matrix<int, Eigen::Dynamic, 5, Eigen::RowMajor> IMatX5_RM; // azimuthal making real typedef Eigen::Matrix<Real, Dynamic, nPEM, Eigen::RowMajor> RMatXN_RM; typedef Eigen::Matrix<Real, Dynamic, nPEM * 3, Eigen::RowMajor> RMatXN3_RM; typedef Eigen::Matrix<Real, Dynamic, nPEM * 6, Eigen::RowMajor> RMatXN6_RM; typedef Eigen::Matrix<Real, Dynamic, nPEM * 9, Eigen::RowMajor> RMatXN9_RM; } #endif /* eigen_element_op_hpp */
# test spark-streaming-local rm -rf /root/test_bmr_spark/output; ${SPARK_HOME}/bin/spark-submit --class WordCount uber-spark_word_count_normal-1.0.0-snapshot.jar file:///root/test_bmr_spark/README file:///root/test_bmr_spark/output # test spark-streaming-yarn hdfs dfs -rmr bos://bmrtest-bj/4c0c80c8-5550-439e-7bce-9c72fe6fdde9/daemons/output.spark; spark-submit --master yarn-cluster --class WordCountYarn uber-spark_word_count_normal-1.0.0-snapshot.jar bos://bmrtest-bj/4c0c80c8-5550-439e-7bce-9c72fe6fdde9/daemons/79e0aabc-818f-4fbf-b487-216851276912/ bos://bmrtest-bj/4c0c80c8-5550-439e-7bce-9c72fe6fdde9/daemons/output.spark #test spark-streaming-yarn with no dependency hdfs dfs -rmr bos://bmrtest-bj/4c0c80c8-5550-439e-7bce-9c72fe6fdde9/daemons/output.spark; spark-submit --master yarn-cluster --class WordCountYarn spark_word_count_normal-1.0.0-snapshot-WordCountYarn-Yarn.jar bos://bmrtest-bj/4c0c80c8-5550-439e-7bce-9c72fe6fdde9/daemons/79e0aabc-818f-4fbf-b487-216851276912/ bos://bmrtest-bj/4c0c80c8-5550-439e-7bce-9c72fe6fdde9/daemons/output.spark
#!/bin/bash set -ex # Wait for docker, else network might not be ready yet while [[ `systemctl status docker | grep active | wc -l` -eq 0 ]] do sleep 2 done # enable CPU manager # kubeadm 1.11 uses a new config method for the kubelet if [ -f /etc/sysconfig/kubelet ]; then # TODO use config file! this is deprecated cat <<EOT >>/etc/sysconfig/kubelet KUBELET_CPUMANAGER_ARGS=--feature-gates=CPUManager=true --cpu-manager-policy=static --kube-reserved=cpu=500m --system-reserved=cpu=500m EOT else cat <<EOT >>/etc/systemd/system/kubelet.service.d/09-kubeadm.conf Environment="KUBELET_CPUMANAGER_ARGS=--feature-gates=CPUManager=true --cpu-manager-policy=static --kube-reserved=cpu=500m --system-reserved=cpu=500m" EOT fi sed -i 's/$KUBELET_EXTRA_ARGS/$KUBELET_EXTRA_ARGS $KUBELET_CPUMANAGER_ARGS/' /etc/systemd/system/kubelet.service.d/10-kubeadm.conf systemctl daemon-reload service kubelet restart kubelet_rc=$? if [[ $kubelet_rc -ne 0 ]]; then rm -rf /var/lib/kubelet/cpu_manager_state service kubelet restart fi kubeadm join --token abcdef.1234567890123456 192.168.66.101:6443 --ignore-preflight-errors=all --discovery-token-unsafe-skip-ca-verification=true
#!/bin/bash # # Licensed to the Apache Software Foundation (ASF) under one # or more contributor license agreements. See the NOTICE file # distributed with this work for additional information # regarding copyright ownership. The ASF licenses this file # to you under the Apache License, Version 2.0 (the # "License"); you may not use this file except in compliance # with the License. You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, # software distributed under the License is distributed on an # "AS IS" BASIS, WITHOUT WARRANTIES OR CONDITIONS OF ANY # KIND, either express or implied. See the License for the # specific language governing permissions and limitations # under the License. # Install some useful/necessary dependencies to make future installs easier sudo apt-get update sudo apt-get install -y make sudo apt-get install -y libblas-dev sudo apt-get install -y liblapack-dev sudo apt-get install -y gfortran sudo apt-get install -y g++ sudo apt-get install -y build-essential sudo apt-get install -y python-dev sudo apt-get install -y ia32-libs --fix-missing sudo apt-get install -y git sudo apt-get install -y vim # GUI related installs sudo apt-get install -y lightdm # XFCE #sudo apt-get install -y xfce4 #sudo apt-get install -y xdg-utils #sudo apt-get install -y eog # Ubuntu Unity sudo apt-get install -y ubuntu-desktop # Use the Easy-OCW Ubuntu install script to get everything # else installed! git clone http://git-wip-us.apache.org/repos/asf/climate.git # Copy the Easy-OCW install script for Ubuntu cp climate/easy-ocw/install-ubuntu-12_04.sh . # Copy the requirements files for conda and pip used by Easy-OCW cp climate/easy-ocw/*.txt . bash install-ubuntu.sh -q # Set symlink for the UI frontend code cd climate/ocw-ui/backend ln -s ../frontend/app app # Cleanup Anaconda and Basemap downloads from the install script cd sudo rm -f Anaconda-1.9.2-Linux-x86_64.sh sudo rm -f basemap-1.0.7.tar.gz sudo rm -rf basemap-1.0.7 mkdir /home/vagrant/Desktop # These links will only work if we're using the Unity desktop. # If you want to use the XFCE desktop you will need to change # the references to 'nautilus' to whatever file browser you # will be using. cat >/home/vagrant/Desktop/climate.desktop <<CODELINK [Desktop Entry] Name=Climate-Code Icon=utilities-terminal Exec=nautilus /home/vagrant/climate Terminal=false Type=Application CODELINK cat >/home/vagrant/Desktop/ui.desktop <<UISTART [Desktop Entry] Name=Climate-UI Icon=utilities-terminal Exec=/home/vagrant/Desktop/.ui.sh Terminal=true Type=Application UISTART cat >/home/vagrant/Desktop/.ui.sh <<UIBOOTUP #!/bin/bash export PATH=/home/vagrant/anaconda/bin:$PATH export PYTHONPATH=/home/vagrant/climate:/home/vagrant/climate/ocw cd ~/climate/ocw-ui/backend && python run_webservices.py UIBOOTUP # It is possible that these commands will need to be rerun once # the desktop environment has been updated. You will notice that # they aren't recognized as shortcuts if you need to rerun these # commands. If all is working properly, you should see terminal # icons with the names listed above (Climate-Code or Climate-UI). chmod +x ~/Desktop/climate.desktop chmod +x ~/Desktop/.ui.sh chmod +x ~/Desktop/ui.desktop
<filename>src/pasa/cbentley/layouter/swing/demo/RunLayouterDemoSwingAbstract.java /* * (c) 2018-2020 <NAME> * This code is licensed under MIT license (see LICENSE.txt for details) */ package pasa.cbentley.layouter.swing.demo; import java.awt.Image; import java.io.IOException; import java.util.List; import javax.management.RuntimeErrorException; import javax.swing.JComponent; import pasa.cbentley.byteobjects.src4.core.ByteObject; import pasa.cbentley.byteobjects.src4.ctx.BOCtx; import pasa.cbentley.core.src4.ctx.IConfigU; import pasa.cbentley.core.src4.interfaces.IPrefs; import pasa.cbentley.core.src4.logging.ILogConfigurator; import pasa.cbentley.core.src5.task.SaveCtxTask; import pasa.cbentley.layouter.src4.engine.PozerFactory; import pasa.cbentley.layouter.src4.engine.SizerFactory; import pasa.cbentley.layouter.src4.engine.Zer2DPozer; import pasa.cbentley.layouter.swing.ctx.SwingLayouterCtx; import pasa.cbentley.layouter.swing.demo.log.LogConfiguratorDemoLayouter; import pasa.cbentley.layouter.swing.engine.JFrameLayouterDemo; import pasa.cbentley.layouter.swing.engine.JPanelLayoutable; import pasa.cbentley.swing.run.RunSwingAbstract; import pasa.cbentley.swing.window.CBentleyFrame; public abstract class RunLayouterDemoSwingAbstract extends RunSwingAbstract { protected final SwingLayouterCtx slc; protected final BOCtx boc; protected PozerFactory pozerFac; protected ByteObject centerCenter; protected ByteObject endToEnd; protected ByteObject topToTop; protected ByteObject pozerBottomToTop; protected ByteObject pozerTopToBottom; protected ByteObject pozerLeftToRight; protected ByteObject pozerRighToRight; protected ByteObject pozerEndtoEnd; protected ByteObject pozerEndtoStart; protected ByteObject pozerCenterToTop; protected ByteObject pozerCenterToLeft; protected ByteObject pozerCenterToRight; protected ByteObject pozerCenterToBot; protected ByteObject pozerYTop; protected ByteObject startToStart; protected ByteObject botToBot; protected Zer2DPozer centerRightParent; protected Zer2DPozer topEndParent; protected Zer2DPozer midEndParent; protected Zer2DPozer botEndParent; protected Zer2DPozer topCenterParent; protected Zer2DPozer midCenterParent; protected Zer2DPozer botCenterParent; protected Zer2DPozer topStartParent; protected Zer2DPozer midStartParent; protected Zer2DPozer botStartParent; private Runnable exitRun; protected SizerFactory sizerFac; protected ByteObject sizerRatio50; protected ByteObject sizerRatio30; protected ByteObject sizerRatio10; protected ByteObject sizerRatio5; protected JPanelLayoutable panel; private String fileName; public RunLayouterDemoSwingAbstract(IConfigU configU) { super(configU); boc = new BOCtx(uc); slc = new SwingLayouterCtx(boc, sc); } public RunLayouterDemoSwingAbstract(SwingLayouterCtx slc) { super(slc.getSwingCtx()); this.boc = slc.getBOC(); this.slc = slc; } public void constructHelpers() { pozerFac = slc.getFactoryPozer(); sizerFac = slc.getSizerFactory(); sizerRatio50 = sizerFac.getSizerRatio100Parent(50); sizerRatio30 = sizerFac.getSizerRatio100Parent(30); sizerRatio10 = sizerFac.getSizerRatio100Parent(10); sizerRatio5 = sizerFac.getSizerRatio100Parent(5); centerCenter = pozerFac.getPozerCenterToCenter(); endToEnd = pozerFac.getPozerBottomRight(); topToTop = pozerFac.getPozerTopToTop(); pozerBottomToTop = pozerFac.getPozerBotToTop(); pozerTopToBottom = pozerFac.getPozerTopToBottom(); pozerLeftToRight = pozerFac.getPozerTopToBottom(); pozerRighToRight = pozerFac.getPozerTopToBottom(); pozerEndtoEnd = pozerFac.getPozerEndToEnd(); pozerEndtoStart = pozerFac.getPozerEndToStart(); pozerCenterToTop = pozerFac.getPozerCenterToTopLeft(); pozerCenterToLeft = pozerCenterToTop; pozerCenterToRight = pozerFac.getPozerCenterToBotRight(); pozerCenterToBot = pozerCenterToRight; pozerYTop = pozerFac.getPozerTopLeft(); startToStart = pozerFac.getPozerStartToStart(); botToBot = pozerFac.getPozerBotToBot(); centerRightParent = new Zer2DPozer(slc, endToEnd, pozerYTop); topEndParent = new Zer2DPozer(slc, endToEnd, topToTop); midEndParent = new Zer2DPozer(slc, endToEnd, centerCenter); botEndParent = new Zer2DPozer(slc, endToEnd, botToBot); topCenterParent = new Zer2DPozer(slc, centerCenter, topToTop); midCenterParent = new Zer2DPozer(slc, centerCenter, centerCenter); botCenterParent = new Zer2DPozer(slc, centerCenter, botToBot); topStartParent = new Zer2DPozer(slc, startToStart, topToTop); midStartParent = new Zer2DPozer(slc, startToStart, centerCenter); botStartParent = new Zer2DPozer(slc, startToStart, botToBot); } public Runnable getExitRun() { return exitRun; } /** * Add * @param exitRun */ public void setExitRun(Runnable exitRun) { if (this.exitRun != null) { //multi run throw new RuntimeException("not implemented"); } this.exitRun = exitRun; } public void cmdExit() { if (exitRun != null) { exitRun.run(); } } /** * Add all the {@link JComponent} that you want * @param panel */ public abstract void buildDemo(JPanelLayoutable panel); protected void addI18n(List<String> list) { } protected void initOutsideUIForPrefs(IPrefs prefs) { //create swing component that supports Drawer //reading from disk the core settings fileName = this.getClass().getSimpleName() + ".state"; try { c5.loadCtxSettingsFromUserHome(fileName); } catch (IOException e) { e.printStackTrace(); } setExitRun(new SaveCtxTask(c5, fileName)); } protected CBentleyFrame initUIThreadInsideSwing() { JFrameLayouterDemo frame = new JFrameLayouterDemo(slc); panel = frame.getPanelLayouter(); this.frame = frame; buildDemo(panel); frame.setTitle(this.getClass().getSimpleName()); Image icon = sc.createImage("/logo_layouter_demo_64.png", ""); frame.setIconImage(icon); //frame.showDefClosePackLocationNull(); //#debug //uc.toDLog().pFlow("Start", frame, RunLayouterDemoSwingAbstract.class, "initUIThreadInsideSwing@line185"); return frame; } public ILogConfigurator toStringGetLoggingConfig() { return new LogConfiguratorDemoLayouter(); } }
<filename>dist/controllers/user.d.ts import { Request, Response } from 'express'; export declare let postLogin: (req: Request<import("express-serve-static-core").ParamsDictionary>, res: Response) => Response; export declare let getUser: (req: Request<import("express-serve-static-core").ParamsDictionary>, res: Response) => Response; export declare let postLogout: (req: Request<import("express-serve-static-core").ParamsDictionary>, res: Response) => void;
package com.threathunter.bordercollie.slot.tool; import java.io.*; import java.util.ArrayList; import java.util.concurrent.CountDownLatch; public class ITLauncher { public static void main(String[] args) throws Exception { String javaHome = System.getProperty("java.home"); final String javaLauncher = new File(javaHome, "bin/java").getPath(); final String classPath = System.getProperty("java.class.path"); final String currentDirectory = System.getProperty("user.dir"); // final String currentClassName = getCurrentClassName(); final String currentClassName = ITMain.class.getCanonicalName(); final String[] processArgs = args; System.out.println("java home = " + javaHome); System.out.println("java launcher = " + javaLauncher); System.out.println("current directory = " + currentDirectory); System.out.println("class name = " + currentClassName); args = new String[]{"-launch"}; if (args.length > 0 && args[0].equals("-launch")) { final ArrayList<String> argumentList = new ArrayList<String>() { { add(javaLauncher); add("-cp"); add(classPath); add(currentClassName); /* for(int i= 1; i<processArgs.length; i++) { add(processArgs[i]); }*/ } }; System.out.println(argumentList); CountDownLatch latch = new CountDownLatch(1); Runnable processLauncher = new Runnable() { @Override public void run() { try { ProcessBuilder processBuilder = new ProcessBuilder(argumentList); processBuilder.redirectErrorStream(true); Process process = processBuilder.start(); StringBuilder processOutput = new StringBuilder(); try (BufferedReader processOutputReader = new BufferedReader( new InputStreamReader(process.getInputStream()));) { String readLine; while ((readLine = processOutputReader.readLine()) != null) { processOutput.append(readLine + System.lineSeparator()); } process.waitFor(); System.out.println(processOutput); latch.countDown(); } } catch (Exception e) { e.printStackTrace(); } } }; new Thread(processLauncher).start(); latch.await(); System.out.println("ITLauncher completed"); } else { FileWriter writer = new FileWriter(File.createTempFile("launch-", ".log", new File(currentDirectory))); writer.write("Launched"); writer.close(); Thread.sleep(10000); } } private static String getCurrentClassName() throws Exception { StackTraceElement[] stackTrace = Thread.currentThread().getStackTrace(); int index = stackTrace.length - 2; StackTraceElement stackTraceElement = stackTrace[index]; Class<?> c = Class.forName(stackTraceElement.getClassName()); return c.getCanonicalName(); } }
import React, { useState, useEffect } from 'react' import { styled } from 'linaria/react' import warning from '@assets/warning.png' import ok from '@assets/ok.png' export const Alert = ({ runEffect, success = false, color, text = success ? 'Успешно' : 'Произошла ошибка' }) => { const [visible, setVisible] = useState(false) useEffect(() => { setVisible(true) runEffect && runEffect() setTimeout(() => { setVisible(false) }, 5000) }, [runEffect]) return ( <Body visible={visible} onClick={() => setVisible(false)} success={success} color={color} > <Img src={success ? ok : warning} alt="" /> <p>{text}</p> </Body> ) } const Body = styled.div` position: absolute; display: ${props => (props.visible ? 'flex' : 'none')}; justify-content: space-between; padding: 0 10px; align-items: center; top: 60px; right: 10px; min-width: 180px; min-height: 40px; border-radius: 3px; background-color: ${props => props.color ? props.color : props.success ? 'var(--green)' : 'var(--red)'}; color: white; font-weight: 400; cursor: pointer; text-align: center; ` const Img = styled.img` width: 20px; height: 20px; margin-right: 10px; `
from pybloom_live import BloomFilter # Common words to store in bloom filter words = ["the", "be", "and", "of", "a", "in", "to", "have", "to", "it", "I", "that", "for", "you", "he", "with", "on", "do", "say", "this", "they", "is", "an", "at", "but", "we", "his", "from", "that", "not", "by", "she", "or", "as", "what", "go", "their", "can", "who", "get", "if", "would", "her", "all", "my", "make", "about", "know"] # Create a bloom filter with a capacity of 100 words and 0.1% error rate bf = BloomFilter(100, 0.1) # Add the words to the bloom filter for word in words: bf.add(word) # Check if some words are in the list or not print("the" in bf) print("the " in bf) print("hello" in bf)
package com.emmanuellmota.metamodel; @GenerateModel public class ImmutableObject { private final String name; public ImmutableObject(String name) { this.name = name; } public String getName() { return name; } }
fn slugify_label(section: &str, label: String) -> String { let section_lower = section.to_lowercase(); let label_modified = label .to_lowercase() .replace(" ", "_") .chars() .filter(|c| c.is_alphanumeric() || *c == '_') .collect::<String>(); format!("{}:{}", section_lower, label_modified) }
package io.cattle.platform.api.parser; import io.cattle.platform.archaius.util.ArchaiusUtil; import io.github.ibuildthecloud.gdapi.request.ApiRequest; import io.github.ibuildthecloud.gdapi.request.parser.DefaultApiRequestParser; import java.io.IOException; import javax.servlet.http.HttpServletRequest; import org.apache.commons.lang3.ArrayUtils; import org.apache.commons.lang3.StringUtils; import com.netflix.config.DynamicBooleanProperty; public class ApiRequestParser extends DefaultApiRequestParser { private static final DynamicBooleanProperty ALLOW_OVERRIDE = ArchaiusUtil.getBoolean("api.allow.client.override"); @Override public boolean isAllowClientOverrideHeaders() { return ALLOW_OVERRIDE.get(); } @Override public boolean parse(ApiRequest apiRequest) throws IOException { HttpServletRequest request = apiRequest.getServletContext().getRequest(); String path = request.getServletPath(); String[] parts = path.split("/"); if (parts.length > 4 && "projects".equalsIgnoreCase(parts[2]) && !"projectMembers".equalsIgnoreCase(parts[4])) { String projectId = parts[3]; apiRequest.setSubContext(String.format("/%s/%s", parts[2], projectId)); String[] newPath = ArrayUtils.addAll(new String[]{"", parts[1]}, ArrayUtils.subarray(parts, 4, Integer.MAX_VALUE)); String servletPath = StringUtils.join(newPath, "/"); request = new ProjectHttpServletRequest(request, projectId, servletPath); apiRequest.getServletContext().setRequest(request); } return super.parse(apiRequest); } }
import './src/marble.scss';
#!/usr/bin/env bash set -e QT_CFG='' BUILD_CONFIRM=0 COMPILE_JOBS=1 MAKEFLAGS_JOBS='' if [[ "$MAKEFLAGS" != "" ]]; then MAKEFLAGS_JOBS=$(echo $MAKEFLAGS | egrep -o '\-j[0-9]+' | egrep -o '[0-9]+') fi if [[ "$MAKEFLAGS_JOBS" != "" ]]; then # user defined number of jobs in MAKEFLAGS, re-use that number COMPILE_JOBS=$MAKEFLAGS_JOBS elif [[ $OSTYPE = darwin* ]]; then # We only support modern Mac machines, they are at least using # hyperthreaded dual-core CPU. COMPILE_JOBS=4 elif [[ $OSTYPE == freebsd* ]]; then COMPILE_JOBS=`sysctl -n hw.ncpu` else CPU_CORES=`grep -c ^processor /proc/cpuinfo` if [[ "$CPU_CORES" -gt 1 ]]; then COMPILE_JOBS=$CPU_CORES fi fi if [[ "$COMPILE_JOBS" -gt 8 ]]; then # Safety net. COMPILE_JOBS=8 fi until [ -z "$1" ]; do case $1 in "--qt-config") shift QT_CFG=" $1" shift;; "--qmake-args") shift QMAKE_ARGS=$1 shift;; "--jobs") shift COMPILE_JOBS=$1 shift;; "--confirm") BUILD_CONFIRM=1 shift;; "--help") echo "Usage: $0 [--qt-config CONFIG] [--jobs NUM]" echo echo " --confirm Silently confirm the build." echo " --qt-config CONFIG Specify extra config options to be used when configuring Qt" echo " --jobs NUM How many parallel compile jobs to use. Defaults to 4." echo exit 0 ;; *) echo "Unrecognised option: $1" exit 1;; esac done if [[ "$BUILD_CONFIRM" -eq 0 ]]; then cat << EOF ---------------------------------------- WARNING ---------------------------------------- Building PhantomJS from source takes a very long time, anywhere from 30 minutes to several hours (depending on the machine configuration). It is recommended to use the premade binary packages on supported operating systems. For details, please go the the web site: http://phantomjs.org/download.html. EOF echo "Do you want to continue (y/n)?" read continue if [[ "$continue" != "y" ]]; then exit 1 fi echo echo fi cd src/qt && ./preconfig.sh --jobs $COMPILE_JOBS --qt-config "$QT_CFG" && cd ../.. src/qt/bin/qmake $QMAKE_ARGS make -j$COMPILE_JOBS
'use strict'; var study = angular.module('study', []); study.controller('StudyCtrl', ['$scope', '$routeParams', '$window', '$location', 'IndexedDb', '$timeout', function ($scope, $routeParams, $window, $location, IndexedDb, $timeout) { $scope.collectionId = $routeParams.id; $scope.cards = []; $scope.card = null; $scope.displayAnswer = false; $scope.onlyAnswer = false; $scope.timer = null; const CARD_GRADE = { 0: 'Again', 1: 'Hard', 3: 'Good', 5: 'Easy' }; $scope.CARD_GRADE = CARD_GRADE; $scope.showAnswer = function () { $scope.displayAnswer = true; $timeout.cancel($scope.countdownRef); }; $scope.evaluateCard = function (grade) { //nastavení dalšího zobazení podle intervalu a stupne odpovědi if ($scope.card.interval < 6) { if (grade >= 3) { var nextShowDay = moment().add(1, "d"); nextShowDay.set({'hour': 0, 'minute': 0, 'second': 0}); $scope.card.nextShow = nextShowDay.toDate(); } else { $scope.card.nextShow = moment().add(5, "m").toDate(); } } else { if (grade == 0) { $scope.card.nextShow = moment().add(5, "m").toDate(); } else { var fraction = 1; if (grade == 3) { fraction = 2 / 3; } else if (grade == 1) { fraction = 1 / 3; } var nextShowDay = moment().add(Math.round($scope.card.interval * fraction), "d"); nextShowDay.set({'hour': 0, 'minute': 0, 'second': 0}); $scope.card.nextShow = nextShowDay.toDate(); } } if (grade >= 3) { if ($scope.card.numberOfIteration == 0) { $scope.card.numberOfIteration = 1; $scope.card.interval = 1; } else if ($scope.card.numberOfIteration == 1) { $scope.card.numberOfIteration = 2; $scope.card.interval = 6; } else { $scope.card.numberOfIteration += 1; $scope.card.interval = Math.round($scope.card.interval * $scope.card.ef); } var newEf = ($scope.card.ef + (0.1 - (5 - grade) * (0.08 + (5 - grade) * 0.02))); $scope.card.ef = ( newEf < 1.3 ) ? 1.3 : newEf; } else { $scope.card.numberOfIteration = 1; $scope.card.interval = 1; } $scope.oldCardDirty = $scope.card.dirty; if (grade == 0) { $scope.card.dirty = "true"; } else { $scope.card.dirty = "false"; } $scope.card.lastShow = moment().toDate(); IndexedDb.update(IndexedDb.STORES.CARD_STORE, $scope.card).then(function () { $scope.updateStatisticsAnswers(grade, $scope.oldCardDirty); $scope.loadStudyCard(); }, function (err) { $window.alert(err); }); }; $scope.speakFrontCardText = function () { meSpeak.loadVoice("js/mespeak/voices/" + $scope.collectionSettings.languageOfVoice + ".json", function () { meSpeak.speak($scope.card.front, { volume: ($scope.collectionSettings.volumeOfVoice / 100), variant: $scope.collectionSettings.typeOfVoice }); }); }; $scope.loadStudyCard = function () { IndexedDb.findCardToStudy($routeParams.id).then( function (data) { $scope.cards = data; if (data.length > 0) { $scope.card = data[0]; if ($scope.collectionSettings.limitTAnswer.toLowerCase() == "true") { $scope.timer = $scope.collectionSettings.maximalAnswerTime; $timeout.cancel($scope.countdownRef); $scope.countdown(); } } else { $timeout.cancel($scope.countdownRef); $location.path('/collections/' + $routeParams.id); } if ($scope.collectionSettings.playVoiceText.toLowerCase() === 'true') { $scope.speakFrontCardText(); } }, function (err) { $window.alert(err); $timeout.cancel($scope.countdownRef); $location.path('/collections/' + $routeParams.id); }); $scope.displayAnswer = false; }; $scope.getNewCards = function (allCards) { var newCards = []; for (var i = 0; i < allCards.length; i++) { if (allCards[i].lastShow == null) { newCards.push(allCards[i]); } } return newCards; }; $scope.getRepeatedCards = function (allCards) { var repeatedCards = []; for (var i = 0; i < allCards.length; i++) { if (allCards[i].dirty == "true") { repeatedCards.push(allCards[i]); } } return repeatedCards; }; $scope.updateStatisticsAnswers = function (grade, dirty) { var today = moment(); today.set({'hour': 0, 'minute': 0, 'second': 0}); today = today.format("YYYY-MM-DD"); //Pro statistiky globalní (bez kolekce) -------------------- IndexedDb.findByProperty(IndexedDb.STORES.STATISTICS_ANSWERS_STORE, 'collectionId', 0).then(function (response) { var record = null; angular.forEach(response, function (value, key) { if (moment(value.day).isSame(today, 'day')) { record = value; } }); if (record === null) { // neexistuje -> vytvořit var newStatisticsToday = { day: today, collectionId: 0, again: 0, hard: 0, good: 0, easy: 0 }; if (dirty == "false") { newStatisticsToday[CARD_GRADE[grade].toLowerCase()] = 1; } IndexedDb.add(IndexedDb.STORES.STATISTICS_ANSWERS_STORE, newStatisticsToday).then(function () { //ok }, function (err) { $window.alert(err); }); } else { //existuje - melo by být jedno -> upravit grade if (dirty == "false") { record[CARD_GRADE[grade].toLowerCase()] += 1; } IndexedDb.update(IndexedDb.STORES.STATISTICS_ANSWERS_STORE, record).then(function () { //ok }, function (err) { $window.alert(err); }); } }, function (err) { $window.alert(err); }); //Pro statistiky kolekci-------------------- IndexedDb.findByProperty(IndexedDb.STORES.STATISTICS_ANSWERS_STORE, 'collectionId', parseInt($scope.collectionId)).then(function (response) { var record = null; angular.forEach(response, function (value, key) { if (moment(value.day).isSame(today, 'day')) { record = value; } }); if (record === null) { // neexistuje -> vytvořit var newStatisticsToday = { day: today, collectionId: parseInt($scope.collectionId), again: 0, hard: 0, good: 0, easy: 0 }; if (dirty == "false") { newStatisticsToday[CARD_GRADE[grade].toLowerCase()] = 1; } IndexedDb.add(IndexedDb.STORES.STATISTICS_ANSWERS_STORE, newStatisticsToday).then(function () { //ok }, function (err) { $window.alert(err); }); } else { //existuje - melo by být jedno -> upravit grade if (dirty == "false") { record[CARD_GRADE[grade].toLowerCase()] += 1; } IndexedDb.update(IndexedDb.STORES.STATISTICS_ANSWERS_STORE, record).then(function () { //ok }, function (err) { $window.alert(err); }); } }, function (err) { $window.alert(err); }); }; $scope.goBack = function() { $timeout.cancel($scope.countdownRef); $location.path('/collections/' + $routeParams.id); }; $scope.countdown = function () { $scope.countdownRef = $timeout(function () { if ($scope.timer <= 0) { $scope.onlyAnswer = true; $timeout(function() { $scope.evaluateCard(0); $scope.timer++; $scope.onlyAnswer = false; }, 1500); } else { $scope.timer--; $scope.countdown(); } }, 1000); }; $scope.init = function () { IndexedDb.open().then(function () { IndexedDb.getById(IndexedDb.STORES.COLLECTION_SETTINGS_STORE, $routeParams.id).then(function (response) { $scope.collectionSettings = response; $scope.loadStudyCard(); meSpeak.loadConfig("js/mespeak/mespeak_config.json"); }, function (err) { $window.alert(err); }); }); } }]);
import random import numpy class QueueManager: def __init__(self, size): self.queue = [element+1 for element in range(size)] def finalState(self): random.shuffle(self.queue) def invalidState(self): for index, element in numpy.ndenumerate(self.queue): if element - (index[0]+1) > 2: self.output = 'Too chaotic' return True return False
fn area_triangle(a: f64, b: f64, c: f64) -> f64 { let s = (a + b + c) / 2.0; (s * (s - a) * (s - b) * (s - c)).sqrt() }
<reponame>dylmeadows/lambdadepot<gh_stars>0 package io.lambdadepot.function.checked; public interface CheckedPredicate0 { boolean test() throws Throwable; }
#!/bin/bash ############################################################################### # This script is used to streamline running E2E tests for Linux. ############################################################################### set -e function clean_up() { print_highlighted_message 'Clean up' echo 'Stop IoT Edge services' systemctl stop iotedge.socket iotedge.mgmt.socket || true systemctl kill iotedge || true systemctl stop iotedge || true echo 'Remove IoT Edge and config file' apt-get purge libiothsm-std --yes || true rm -rf /var/lib/iotedge/ rm -rf /var/run/iotedge/ rm -rf /etc/iotedge/config.yaml if [ "$CLEAN_ALL" = '1' ]; then echo 'Prune docker system' docker system prune -af --volumes || true else echo 'Remove docker containers' docker rm -f $(docker ps -aq) || true fi } function create_iotedge_service_config { print_highlighted_message 'Create IoT Edge service config' mkdir /etc/systemd/system/iotedge.service.d/ || true bash -c "echo '[Service] Environment=IOTEDGE_LOG=edgelet=debug' > /etc/systemd/system/iotedge.service.d/override.conf" } function set_certificate_generation_tools_dir() { if [[ -z $CERT_SCRIPT_DIR ]]; then CERT_SCRIPT_DIR="$E2E_TEST_DIR/artifacts/core-linux/CACertificates" fi } function get_image_architecture_label() { local arch arch="$(uname -m)" case "$arch" in 'x86_64' ) image_architecture_label='amd64';; 'armv7l' ) image_architecture_label='arm32v7';; 'aarch64' ) image_architecture_label='arm64v8';; *) print_error "Unsupported OS architecture: $arch"; exit 1;; esac } function get_iotedge_quickstart_artifact_file() { local path if [ "$image_architecture_label" = 'amd64' ]; then path="$E2E_TEST_DIR/artifacts/core-linux/IotEdgeQuickstart.linux-x64.tar.gz" elif [ "$image_architecture_label" = 'arm64v8' ]; then path="$E2E_TEST_DIR/artifacts/core-linux/IotEdgeQuickstart.linux-arm64.tar.gz" else path="$E2E_TEST_DIR/artifacts/core-linux/IotEdgeQuickstart.linux-arm.tar.gz" fi echo "$path" } function get_iotedged_artifact_folder() { local path if [ "$image_architecture_label" = 'amd64' ]; then path="$E2E_TEST_DIR/artifacts/iotedged-ubuntu16.04-amd64" elif [ "$image_architecture_label" = 'arm64v8' ]; then path="$E2E_TEST_DIR/artifacts/iotedged-ubuntu18.04-aarch64" else path="$E2E_TEST_DIR/artifacts/iotedged-debian9-arm32v7" fi echo "$path" } function get_leafdevice_artifact_file() { local path if [ "$image_architecture_label" = 'amd64' ]; then path="$E2E_TEST_DIR/artifacts/core-linux/LeafDevice.linux-x64.tar.gz" elif [ "$image_architecture_label" = 'arm64v8' ]; then path="$E2E_TEST_DIR/artifacts/core-linux/LeafDevice.linux-arm64.tar.gz" else path="$E2E_TEST_DIR/artifacts/core-linux/LeafDevice.linux-arm.tar.gz" fi echo "$path" } function get_long_haul_deployment_artifact_file() { local path path="$E2E_TEST_DIR/artifacts/core-linux/e2e_deployment_files/long_haul_deployment.template.json" echo "$path" } function prepare_test_from_artifacts() { print_highlighted_message 'Prepare test from artifacts' echo 'Remove working folder' rm -rf "$working_folder" mkdir -p "$working_folder" declare -a pkg_list=( $iotedged_artifact_folder/*.deb ) iotedge_package="${pkg_list[*]}" echo "iotedge_package=$iotedge_package" echo 'Extract quickstart to working folder' mkdir -p "$quickstart_working_folder" tar -C "$quickstart_working_folder" -xzf "$iotedge_quickstart_artifact_file" if [[ "${TEST_NAME,,}" == 'quickstartcerts' ]]; then echo 'Extract leaf device to working folder' mkdir -p "$leafdevice_working_folder" tar -C "$leafdevice_working_folder" -xzf "$leafdevice_artifact_file" fi if [[ "${TEST_NAME,,}" == directmethod* ]] || [[ "${TEST_NAME,,}" == 'longhaul' ]] || [[ "${TEST_NAME,,}" == 'stress' ]] || [[ "${TEST_NAME,,}" == 'tempfilter' ]] || [[ "${TEST_NAME,,}" == 'tempfilterfunctions' ]]; then case "${TEST_NAME,,}" in directmethod*) echo "Copy deployment file from $dm_module_to_module_deployment_artifact_file" cp "$dm_module_to_module_deployment_artifact_file" "$deployment_working_file" case "${TEST_NAME,,}" in 'directmethodamqp') sed -i -e "s@<UpstreamProtocol>@Amqp@g" "$deployment_working_file" sed -i -e "s@<ClientTransportType>@Amqp_Tcp_Only@g" "$deployment_working_file";; 'directmethodamqpmqtt') sed -i -e "s@<UpstreamProtocol>@Amqp@g" "$deployment_working_file" sed -i -e "s@<ClientTransportType>@Mqtt_Tcp_Only@g" "$deployment_working_file";; 'directmethodamqpws') sed -i -e "s@<UpstreamProtocol>@Amqpws@g" "$deployment_working_file" sed -i -e "s@<ClientTransportType>@Amqp_WebSocket_Only@g" "$deployment_working_file";; 'directmethodmqtt') sed -i -e "s@<UpstreamProtocol>@Mqtt@g" "$deployment_working_file" sed -i -e "s@<ClientTransportType>@Mqtt_Tcp_Only@g" "$deployment_working_file";; 'directmethodmqttamqp') sed -i -e "s@<UpstreamProtocol>@Mqtt@g" "$deployment_working_file" sed -i -e "s@<ClientTransportType>@Amqp_Tcp_Only@g" "$deployment_working_file";; 'directmethodmqttws') sed -i -e "s@<UpstreamProtocol>@MqttWs@g" "$deployment_working_file" sed -i -e "s@<ClientTransportType>@Mqtt_WebSocket_Only@g" "$deployment_working_file";; esac;; 'longhaul' | 'stress') if [[ "${TEST_NAME,,}" == 'longhaul' ]]; then echo "Copy deployment file from $long_haul_deployment_artifact_file" cp "$long_haul_deployment_artifact_file" "$deployment_working_file" sed -i -e "s@<DesiredModulesToRestartCSV>@$DESIRED_MODULES_TO_RESTART_CSV@g" "$deployment_working_file" sed -i -e "s@<RestartIntervalInMins>@$RESTART_INTERVAL_IN_MINS@g" "$deployment_working_file" sed -i -e "s@<ServiceClientConnectionString>@$IOTHUB_CONNECTION_STRING@g" "$deployment_working_file" else echo "Copy deployment file from $stress_deployment_artifact_file" cp "$stress_deployment_artifact_file" "$deployment_working_file" sed -i -e "s@<LoadGen1.TransportType>@$LOADGEN1_TRANSPORT_TYPE@g" "$deployment_working_file" sed -i -e "s@<LoadGen2.TransportType>@$LOADGEN2_TRANSPORT_TYPE@g" "$deployment_working_file" sed -i -e "s@<LoadGen3.TransportType>@$LOADGEN3_TRANSPORT_TYPE@g" "$deployment_working_file" sed -i -e "s@<LoadGen4.TransportType>@$LOADGEN4_TRANSPORT_TYPE@g" "$deployment_working_file" sed -i -e "s@<amqpSettings__enabled>@$AMQP_SETTINGS_ENABLED@g" "$deployment_working_file" sed -i -e "s@<mqttSettings__enabled>@$MQTT_SETTINGS_ENABLED@g" "$deployment_working_file" fi local escapedSnitchAlertUrl local escapedBuildId sed -i -e "s@<Analyzer.ConsumerGroupId>@$EVENT_HUB_CONSUMER_GROUP_ID@g" "$deployment_working_file" sed -i -e "s@<Analyzer.EventHubConnectionString>@$EVENTHUB_CONNECTION_STRING@g" "$deployment_working_file" sed -i -e "s@<Analyzer.LogAnalyticsEnabled>@$LOG_ANALYTICS_ENABLED@g" "$deployment_working_file" sed -i -e "s@<Analyzer.LogAnalyticsLogType>@$LOG_ANALYTICS_LOG_TYPE@g" "$deployment_working_file" sed -i -e "s@<LogAnalyticsWorkspaceId>@$LOG_ANALYTICS_WORKSPACE_ID@g" "$deployment_working_file" sed -i -e "s@<LogAnalyticsSharedKey>@$LOG_ANALYTICS_SHARED_KEY@g" "$deployment_working_file" sed -i -e "s@<LoadGen.MessageFrequency>@$LOADGEN_MESSAGE_FREQUENCY@g" "$deployment_working_file" escapedSnitchAlertUrl="${SNITCH_ALERT_URL//&/\\&}" escapedBuildId="${ARTIFACT_IMAGE_BUILD_NUMBER//./}" sed -i -e "s@<Snitch.AlertUrl>@$escapedSnitchAlertUrl@g" "$deployment_working_file" sed -i -e "s@<Snitch.BuildNumber>@$SNITCH_BUILD_NUMBER@g" "$deployment_working_file" sed -i -e "s@<Snitch.BuildId>@$RELEASE_LABEL-$image_architecture_label-linux-$escapedBuildId@g" "$deployment_working_file" sed -i -e "s@<Snitch.ReportingIntervalInSecs>@$SNITCH_REPORTING_INTERVAL_IN_SECS@g" "$deployment_working_file" sed -i -e "s@<Snitch.StorageAccount>@$SNITCH_STORAGE_ACCOUNT@g" "$deployment_working_file" sed -i -e "s@<Snitch.StorageMasterKey>@$SNITCH_STORAGE_MASTER_KEY@g" "$deployment_working_file" sed -i -e "s@<Snitch.TestDurationInSecs>@$SNITCH_TEST_DURATION_IN_SECS@g" "$deployment_working_file";; 'tempfilter') echo "Copy deployment file from $module_to_module_deployment_artifact_file" cp "$module_to_module_deployment_artifact_file" "$deployment_working_file";; 'tempfilterfunctions') echo "Copy deployment file from $module_to_functions_deployment_artifact_file" cp "$module_to_functions_deployment_artifact_file" "$deployment_working_file";; esac sed -i -e "s@<Architecture>@$image_architecture_label@g" "$deployment_working_file" sed -i -e "s/<Build.BuildNumber>/$ARTIFACT_IMAGE_BUILD_NUMBER/g" "$deployment_working_file" sed -i -e "s@<CR.Username>@$CONTAINER_REGISTRY_USERNAME@g" "$deployment_working_file" sed -i -e "s@<CR.Password>@$CONTAINER_REGISTRY_PASSWORD@g" "$deployment_working_file" sed -i -e "s@<Container_Registry>@$CONTAINER_REGISTRY@g" "$deployment_working_file" fi } function print_error() { local message=$1 local red='\033[0;31m' local color_reset='\033[0m' echo -e "${red}$message${color_reset}" } function print_highlighted_message() { local message=$1 local cyan='\033[0;36m' local color_reset='\033[0m' echo -e "${cyan}$message${color_reset}" } function print_logs() { local ret=$1 local test_end_time=$2 local elapsed_seconds=$3 elapsed_time="$(TZ=UTC0 printf '%(%H:%M:%S)T\n' "$elapsed_seconds")" print_highlighted_message "Test completed at $test_end_time, took $elapsed_time." if (( ret < 1 )); then return; fi print_highlighted_message 'Print logs' print_highlighted_message 'LOGS FROM IOTEDGED' journalctl -u iotedge -u docker --since "$test_start_time" --no-pager || true print_highlighted_message 'EDGE AGENT LOGS' docker logs edgeAgent || true print_highlighted_message 'EDGE HUB LOGS' docker logs edgeHub || true if [[ "${TEST_NAME,,}" == 'tempsensor' ]]; then print_highlighted_message 'TEMP SENSOR LOGS' docker logs tempSensor || true fi if [[ "${TEST_NAME,,}" == 'tempfilter' ]]; then print_highlighted_message 'TEMP FILTER LOGS' docker logs tempFilter || true fi if [[ "${TEST_NAME,,}" == 'tempfilterfunctions' ]]; then print_highlighted_message 'TEMP FILTER FUNCTIONS LOGS' docker logs tempFilterFunctions || true fi if [[ "${TEST_NAME,,}" == directmethod* ]]; then print_highlighted_message 'DIRECT MTEHOD SENDER LOGS' docker logs DirectMethodSender || true print_highlighted_message 'DIRECT MTEHOD RECEIVER LOGS' docker logs DirectMethodReceiver || true fi } function process_args() { print_highlighted_message 'Process arguments' saveNextArg=0 for arg in "$@" do if [ $saveNextArg -eq 1 ]; then E2E_TEST_DIR="$arg" saveNextArg=0 elif [ $saveNextArg -eq 2 ]; then RELEASE_LABEL="$arg" saveNextArg=0 elif [ $saveNextArg -eq 3 ]; then ARTIFACT_IMAGE_BUILD_NUMBER="$arg" saveNextArg=0 elif [ $saveNextArg -eq 4 ]; then TEST_NAME="$arg" saveNextArg=0 elif [ $saveNextArg -eq 5 ]; then CONTAINER_REGISTRY="$arg" saveNextArg=0 elif [ $saveNextArg -eq 6 ]; then CONTAINER_REGISTRY_USERNAME="$arg" saveNextArg=0 elif [ $saveNextArg -eq 7 ]; then CONTAINER_REGISTRY_PASSWORD="$arg" saveNextArg=0 elif [ $saveNextArg -eq 8 ]; then IOTHUB_CONNECTION_STRING="$arg" saveNextArg=0 elif [ $saveNextArg -eq 9 ]; then EVENTHUB_CONNECTION_STRING="$arg" saveNextArg=0 elif [ $saveNextArg -eq 10 ]; then LOADGEN_MESSAGE_FREQUENCY="$arg" saveNextArg=0 elif [ $saveNextArg -eq 11 ]; then SNITCH_ALERT_URL="$arg" saveNextArg=0 elif [ $saveNextArg -eq 12 ]; then SNITCH_BUILD_NUMBER="$arg" saveNextArg=0 elif [ $saveNextArg -eq 13 ]; then SNITCH_REPORTING_INTERVAL_IN_SECS="$arg" saveNextArg=0 elif [ $saveNextArg -eq 14 ]; then SNITCH_STORAGE_ACCOUNT="$arg" saveNextArg=0 elif [ $saveNextArg -eq 15 ]; then SNITCH_STORAGE_MASTER_KEY="$arg" saveNextArg=0 elif [ $saveNextArg -eq 16 ]; then SNITCH_TEST_DURATION_IN_SECS="$arg" saveNextArg=0 elif [ $saveNextArg -eq 17 ]; then LOADGEN1_TRANSPORT_TYPE="$arg" saveNextArg=0 elif [ $saveNextArg -eq 18 ]; then LOADGEN2_TRANSPORT_TYPE="$arg" saveNextArg=0 elif [ $saveNextArg -eq 19 ]; then LOADGEN3_TRANSPORT_TYPE="$arg" saveNextArg=0 elif [ $saveNextArg -eq 20 ]; then LOADGEN4_TRANSPORT_TYPE="$arg" saveNextArg=0 elif [ $saveNextArg -eq 21 ]; then AMQP_SETTINGS_ENABLED="$arg" saveNextArg=0 elif [ $saveNextArg -eq 22 ]; then MQTT_SETTINGS_ENABLED="$arg" saveNextArg=0 elif [ $saveNextArg -eq 23 ]; then CERT_SCRIPT_DIR="$arg" saveNextArg=0 elif [ $saveNextArg -eq 24 ]; then ROOT_CA_CERT_PATH="$arg" INSTALL_CA_CERT=1 saveNextArg=0 elif [ $saveNextArg -eq 25 ]; then ROOT_CA_KEY_PATH="$arg" INSTALL_CA_CERT=1 saveNextArg=0 elif [ $saveNextArg -eq 26 ]; then ROOT_CA_PASSWORD="$arg" INSTALL_CA_CERT=1 saveNextArg=0 elif [ $saveNextArg -eq 27 ]; then DPS_SCOPE_ID="$arg" saveNextArg=0 elif [ $saveNextArg -eq 28 ]; then DPS_MASTER_SYMMETRIC_KEY="$arg" saveNextArg=0 elif [ $saveNextArg -eq 29 ]; then EVENT_HUB_CONSUMER_GROUP_ID="$arg" saveNextArg=0 elif [ $saveNextArg -eq 30 ]; then DESIRED_MODULES_TO_RESTART_CSV="$arg" saveNextArg=0 elif [ $saveNextArg -eq 31 ]; then RESTART_INTERVAL_IN_MINS="$arg" saveNextArg=0; elif [ $saveNextArg -eq 32 ]; then LOG_ANALYTICS_ENABLED="$arg" saveNextArg=0 elif [ $saveNextArg -eq 33 ]; then LOG_ANALYTICS_WORKSPACE_ID="$arg" saveNextArg=0 elif [ $saveNextArg -eq 34 ]; then LOG_ANALYTICS_SHARED_KEY="$arg" saveNextArg=0 elif [ $saveNextArg -eq 35 ]; then LOG_ANALYTICS_LOG_TYPE="$arg" saveNextArg=0 else case "$arg" in '-h' | '--help' ) usage;; '-testDir' ) saveNextArg=1;; '-releaseLabel' ) saveNextArg=2;; '-artifactImageBuildNumber' ) saveNextArg=3;; '-testName' ) saveNextArg=4;; '-containerRegistry' ) saveNextArg=5;; '-containerRegistryUsername' ) saveNextArg=6;; '-containerRegistryPassword' ) saveNextArg=7;; '-iotHubConnectionString' ) saveNextArg=8;; '-eventHubConnectionString' ) saveNextArg=9;; '-loadGenMessageFrequency' ) saveNextArg=10;; '-snitchAlertUrl' ) saveNextArg=11;; '-snitchBuildNumber' ) saveNextArg=12;; '-snitchReportingIntervalInSecs' ) saveNextArg=13;; '-snitchStorageAccount' ) saveNextArg=14;; '-snitchStorageMasterKey' ) saveNextArg=15;; '-snitchTestDurationInSecs' ) saveNextArg=16;; '-loadGen1TransportType' ) saveNextArg=17;; '-loadGen2TransportType' ) saveNextArg=18;; '-loadGen3TransportType' ) saveNextArg=19;; '-loadGen4TransportType' ) saveNextArg=20;; '-amqpSettingsEnabled' ) saveNextArg=21;; '-mqttSettingsEnabled' ) saveNextArg=22;; '-certScriptDir' ) saveNextArg=23;; '-installRootCACertPath' ) saveNextArg=24;; '-installRootCAKeyPath' ) saveNextArg=25;; '-installRootCAKeyPassword' ) saveNextArg=26;; '-dpsScopeId' ) saveNextArg=27;; '-dpsMasterSymmetricKey' ) saveNextArg=28;; '-eventHubConsumerGroupId' ) saveNextArg=29;; '-desiredModulesToRestartCSV' ) saveNextArg=30;; '-restartIntervalInMins' ) saveNextArg=31;; '-logAnalyticsEnabled' ) saveNextArg=32;; '-logAnalyticsWorkspaceId' ) saveNextArg=33;; '-logAnalyticsSharedKey' ) saveNextArg=34;; '-logAnalyticsLogType' ) saveNextArg=35;; '-cleanAll' ) CLEAN_ALL=1;; * ) usage;; esac fi done # Required parameters [[ -z "$RELEASE_LABEL" ]] && { print_error 'Release label is required.'; exit 1; } [[ -z "$ARTIFACT_IMAGE_BUILD_NUMBER" ]] && { print_error 'Artifact image build number is required'; exit 1; } [[ -z "$TEST_NAME" ]] && { print_error 'Test name is required'; exit 1; } [[ -z "$CONTAINER_REGISTRY_USERNAME" ]] && { print_error 'Container registry username is required'; exit 1; } [[ -z "$CONTAINER_REGISTRY_PASSWORD" ]] && { print_error 'Container registry password is required'; exit 1; } [[ -z "$IOTHUB_CONNECTION_STRING" ]] && { print_error 'IoT hub connection string is required'; exit 1; } [[ -z "$EVENTHUB_CONNECTION_STRING" ]] && { print_error 'Event hub connection string is required'; exit 1; } [[ -z "$LOG_ANALYTICS_ENABLED" ]] && { LOG_ANALYTICS_ENABLED="false"; } [[ "$LOG_ANALYTICS_ENABLED" == true ]] && \ { [[ -z "$LOG_ANALYTICS_WORKSPACE_ID" ]] && { print_error 'Log Analytics Workspace ID is required'; exit 1; }; \ [[ -z "$LOG_ANALYTICS_SHARED_KEY" ]] && { print_error 'Log Analytics secret is required'; exit 1; }; \ [[ -z "$LOG_ANALYTICS_LOG_TYPE" ]] && { print_error 'Log Analytics Log Type is required'; exit 1; }; } echo 'Required parameters are provided' } function get_hash() { # TODO: testHelper.sh needs to be shared across build pipelines local length=$1 local hash=$(cat /dev/urandom | tr -dc 'a-zA-Z0-9' | head -c $length) echo "$hash" } function run_all_tests() { local funcRet=0 local testRet=0 TEST_NAME='DirectMethodAmqp' run_directmethodamqp_test && funcRet=$? || funcRet=$? TEST_NAME='DirectMethodAmqpMqtt' run_directmethodamqpmqtt_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DirectMethodAmqpws' run_directmethodamqpws_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DirectMethodMqtt' run_directmethodmqtt_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DirectMethodMqttAmqp' run_directmethodmqttamqp_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DirectMethodMqttws' run_directmethodmqttws_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DpsSymmetricKeyProvisioning' run_dps_provisioning_test "SymmetricKey" && funcRet=$? || funcRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DpsTpmProvisioning' run_dps_provisioning_test "Tpm" && funcRet=$? || funcRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='DpsX509Provisioning' run_dps_provisioning_test "X509" && funcRet=$? || funcRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='QuickstartCerts' run_quickstartcerts_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='TempFilter' run_tempfilter_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='TempFilterFunctions' run_tempfilterfunctions_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi TEST_NAME='TempSensor' run_tempsensor_test && testRet=$? || testRet=$? if [ $funcRet -eq 0 ]; then funcRet=$testRet; fi return $funcRet } function run_directmethod_test() { SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" \ --verify-data-from-module "DirectMethodSender" \ -l "$deployment_working_file" && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_directmethodamqp_test() { print_highlighted_message "Run DirectMethod test with Amqp upstream protocol and Amqp client transport type for $image_architecture_label" test_setup device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DMAmqp" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DirectMethod test with Amqp upstream protocol and Amqp client transport type on '$device_id' started at $test_start_time" run_directmethod_test && ret=$? || ret=$? return $ret } function run_directmethodamqpmqtt_test() { print_highlighted_message "Run DirectMethod test with Amqp upstream protocol and Mqtt client transport type for $image_architecture_label" test_setup device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DMAmqpMqtt" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DirectMethod test with Amqp upstream protocol and Mqtt client transport type on '$device_id' started at $test_start_time" run_directmethod_test && ret=$? || ret=$? return $ret } function run_directmethodamqpws_test() { print_highlighted_message "Run DirectMethod test with AmqpWs upstream protocol and AmqpWs client transport type for $image_architecture_label" test_setup device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DMAmqpws" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DirectMethod test with AmqpWs upstream protocol and AmqpWs client transport type on '$device_id' started at $test_start_time" run_directmethod_test && ret=$? || ret=$? return $ret } function run_directmethodmqtt_test() { print_highlighted_message "Run DirectMethod test with Mqtt upstream protocol and Mqtt client transport type for $image_architecture_label" test_setup device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DMMqtt" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DirectMethod test with Mqtt upstream protocol and Mqtt client transport type on '$device_id' started at $test_start_time" run_directmethod_test && ret=$? || ret=$? return $ret } function run_directmethodmqttamqp_test() { print_highlighted_message "Run DirectMethod test with Mqtt upstream protocol and Amqp client transport type for $image_architecture_label" test_setup device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DMMqttAmqp" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DirectMethod test with Mqtt upstream protocol and Amqp client transport type on '$device_id' started at $test_start_time" run_directmethod_test && ret=$? || ret=$? return $ret } function run_directmethodmqttws_test() { print_highlighted_message "Run DirectMethod test with MqttWs upstream protocol and MqttWs client transport type for $image_architecture_label" test_setup device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DMMqttws" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DirectMethod test with MqttWs upstream protocol and MqttWs client transport type on '$device_id' started at $test_start_time" run_directmethod_test && ret=$? || ret=$? return $ret } function run_dps_provisioning_test() { local provisioning_type="${1}" print_highlighted_message "Run DPS provisioning test using $provisioning_type for $image_architecture_label" test_setup local registration_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-DPS-$provisioning_type" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run DPS provisioning test $provisioning_type for registration id '$registration_id' started at $test_start_time" local dps_command_flags="" if [[ $provisioning_type == "SymmetricKey" ]]; then dps_command_flags="--dps-scope-id=$DPS_SCOPE_ID \ --dps-registration-id=$registration_id \ --dps-master-symmetric-key=$DPS_MASTER_SYMMETRIC_KEY" elif [[ $provisioning_type == "Tpm" ]]; then dps_command_flags="--dps-scope-id=$DPS_SCOPE_ID \ --dps-registration-id=$registration_id" else # x.509 provisioning # generate the edge device identity primary certificate and key FORCE_NO_PROD_WARNING="true" ${CERT_SCRIPT_DIR}/certGen.sh create_device_certificate "${registration_id}" local edge_device_id_cert=$(readlink -f ${CERT_SCRIPT_DIR}/certs/iot-device-${registration_id}-full-chain.cert.pem) local edge_device_id_key=$(readlink -f ${CERT_SCRIPT_DIR}/private/iot-device-${registration_id}.key.pem) dps_command_flags="--dps-scope-id=$DPS_SCOPE_ID \ --device_identity_pk=$edge_device_id_key \ --device_identity_cert=$edge_device_id_cert" fi SECONDS=0 local ret=0 # note the registration id is the expected device id to be provisioned by DPS "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$registration_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ -tw "$E2E_TEST_DIR/artifacts/core-linux/e2e_test_files/twin_test_tempSensor.json" \ --optimize_for_performance="$optimize_for_performance" \ $dps_command_flags \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_longhaul_test() { print_highlighted_message "Run Long Haul test for $image_architecture_label" test_setup local device_id="$RELEASE_LABEL-Linux-$image_architecture_label-longhaul-$(get_hash 8)" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run Long Haul test with -d '$device_id' started at $test_start_time" SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" \ --leave-running=All \ -l "$deployment_working_file" \ --runtime-log-level "Info" \ --no-verify && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_quickstartcerts_test() { print_highlighted_message "Run Quickstart Certs test for $image_architecture_label" test_setup local device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-QuickstartCerts" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run Quickstart Certs test with -d '$device_id' started at $test_start_time" SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "doesNotNeed" \ -n "$(hostname)" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" \ --leave-running=Core \ --optimize_for_performance="$optimize_for_performance" \ --no-verify && ret=$? || ret=$? declare -a certs=( /var/lib/iotedge/hsm/certs/edge_owner_ca*.pem ) echo "cert: ${certs[0]}" # Workaround for multiple certificates in the x509store - remove this after quick start certs have Authority Key Identifier rm -rf ~/.dotnet/corefx/cryptography/x509stores/root/ "$leafdevice_working_folder/LeafDevice" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -d "$device_id-leaf" \ -ct "${certs[0]}" \ -ed "$(hostname)" && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_stress_test() { print_highlighted_message "Run Stress test for $image_architecture_label" test_setup local device_id="$RELEASE_LABEL-Linux-$image_architecture_label-stress-$(get_hash 8)" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run Stress test with -d '$device_id' started at $test_start_time" SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "doesNotNeed" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" \ --leave-running=All \ -l "$deployment_working_file" \ --runtime-log-level "Info" \ --no-verify && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_tempfilter_test() { print_highlighted_message "Run TempFilter test for $image_architecture_label" test_setup local device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-tempFilter" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run TempFilter test with -d '$device_id' started at $test_start_time" SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ --verify-data-from-module "tempFilter" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" \ -l "$deployment_working_file" && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_tempfilterfunctions_test() { print_highlighted_message "Run TempFilterFunctions test for $image_architecture_label" test_setup local device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-tempFilterFunc" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run TempFilterFunctions test with -d '$device_id' started at $test_start_time" SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ --verify-data-from-module "tempFilterFunctions" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" \ -l "$deployment_working_file" && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_tempsensor_test() { print_highlighted_message "Run TempSensor test for $image_architecture_label" test_setup local device_id="e2e-$RELEASE_LABEL-Linux-$image_architecture_label-tempSensor" test_start_time="$(date '+%Y-%m-%d %H:%M:%S')" print_highlighted_message "Run TempSensor test with -d '$device_id' started at $test_start_time" SECONDS=0 local ret=0 "$quickstart_working_folder/IotEdgeQuickstart" \ -d "$device_id" \ -a "$iotedge_package" \ -c "$IOTHUB_CONNECTION_STRING" \ -e "$EVENTHUB_CONNECTION_STRING" \ -r "$CONTAINER_REGISTRY" \ -u "$CONTAINER_REGISTRY_USERNAME" \ -p "$CONTAINER_REGISTRY_PASSWORD" \ -n "$(hostname)" \ -tw "$E2E_TEST_DIR/artifacts/core-linux/e2e_test_files/twin_test_tempSensor.json" \ --optimize_for_performance="$optimize_for_performance" \ -t "$ARTIFACT_IMAGE_BUILD_NUMBER-linux-$image_architecture_label" && ret=$? || ret=$? local elapsed_seconds=$SECONDS test_end_time="$(date '+%Y-%m-%d %H:%M:%S')" print_logs $ret "$test_end_time" $elapsed_seconds return $ret } function run_test() { if [[ $INSTALL_CA_CERT -eq 1 ]]; then set_certificate_generation_tools_dir [[ -z "$CERT_SCRIPT_DIR" ]] && { print_error 'Certificate script dir is required'; exit 1; } [[ ! -d "$CERT_SCRIPT_DIR" ]] && { print_error 'Certificate script dir is invalid'; exit 1; } FORCE_NO_PROD_WARNING="true" ${CERT_SCRIPT_DIR}/certGen.sh install_root_ca_from_files ${ROOT_CA_CERT_PATH} ${ROOT_CA_KEY_PATH} ${ROOT_CA_PASSWORD} fi local ret=0 case "${TEST_NAME,,}" in 'all') run_all_tests && ret=$? || ret=$?;; 'directmethodamqp') run_directmethodamqp_test && ret=$? || ret=$?;; 'directmethodamqpmqtt') run_directmethodamqpmqtt_test && ret=$? || ret=$?;; 'directmethodamqpws') run_directmethodamqpws_test && ret=$? || ret=$?;; 'directmethodmqtt') run_directmethodmqtt_test && ret=$? || ret=$?;; 'directmethodmqttamqp') run_directmethodmqttamqp_test && ret=$? || ret=$?;; 'directmethodmqttws') run_directmethodmqttws_test && ret=$? || ret=$?;; 'dpssymmetrickeyprovisioning') run_dps_provisioning_test "SymmetricKey" && ret=$? || ret=$?;; 'dpstpmprovisioning') run_dps_provisioning_test "Tpm" && ret=$? || ret=$?;; 'dpsx509provisioning') run_dps_provisioning_test "X509" && ret=$? || ret=$?;; 'quickstartcerts') run_quickstartcerts_test && ret=$? || ret=$?;; 'longhaul') run_longhaul_test && ret=$? || ret=$?;; 'stress') run_stress_test && ret=$? || ret=$?;; 'tempfilter') run_tempfilter_test && ret=$? || ret=$?;; 'tempfilterfunctions') run_tempfilterfunctions_test && ret=$? || ret=$?;; 'tempsensor') run_tempsensor_test && ret=$? || ret=$?;; *) print_highlighted_message "Can't find any test with name '$TEST_NAME'";; esac echo "Test exit with result code $ret" exit $ret } function test_setup() { validate_test_parameters clean_up prepare_test_from_artifacts create_iotedge_service_config } function validate_test_parameters() { print_highlighted_message "Validate test parameters for $TEST_NAME" local required_files=() local required_folders=() required_files+=("$iotedge_quickstart_artifact_file") required_folders+=("$iotedged_artifact_folder") case "${TEST_NAME,,}" in 'tempsensor') required_files+=($twin_testfile_artifact_file);; 'tempfilter') required_files+=($module_to_module_deployment_artifact_file);; 'tempfilterfunctions') required_files+=($module_to_functions_deployment_artifact_file);; 'longhaul') required_files+=($long_haul_deployment_artifact_file);; 'quickstartcerts') required_files+=($leafdevice_artifact_file);; 'stress') required_files+=($stress_deployment_artifact_file);; esac if [[ "${TEST_NAME,,}" == directmethod* ]]; then required_files+=($dm_module_to_module_deployment_artifact_file) fi local error=0 for f in "${required_files[@]}" do if [ ! -f "$f" ]; then print_error "Required file, $f doesn't exist." ((error++)) fi done for d in "${required_folders[@]}" do if [ ! -d "$d" ]; then print_error "Required directory, $d doesn't exist." ((error++)) fi done if [[ "${TEST_NAME,,}" == "longhaul" ]] || [[ "${TEST_NAME,,}" == "stress" ]]; then if [[ -z "$SNITCH_ALERT_URL" ]]; then print_error "Required snitch alert URL." ((error++)) fi if [[ -z "$SNITCH_STORAGE_ACCOUNT" ]]; then print_error "Required snitch storage account." ((error++)) fi if [[ -z "$SNITCH_STORAGE_MASTER_KEY" ]]; then print_error "Required snitch storage master key." ((error++)) fi fi if (( error > 0 )); then exit 1 fi } function usage() { echo "$SCRIPT_NAME [options]" echo '' echo 'options' echo ' -testDir Path of E2E test directory which contains artifacts and certs folders; defaul to current directory.' echo ' -releaseLabel Release label can be uniquely identify the build (e.g <ReleaseName>-<ReleaseAttempt>); which is used as part of Edge device name.' echo ' -testName Name of E2E test to be run.' echo " Values are 'All', 'DirectMethodAmqp', 'DirectMethodAmqpMqtt', 'DirectMethodAmqpWs', 'DirectMethodMqtt', 'DirectMethodMqttAmqp', " echo " 'DirectMethodMqttWs', 'LongHaul', 'QuickstartCerts', 'Stress', 'TempFilter', 'TempFilterFunctions', 'TempSensor'" echo " 'DpsSymmetricKeyProvisioning', 'DpsTpmProvisioning', 'DpsX509Provisioning'" echo " 'LongHaul', 'QuickstartCerts', 'Stress', 'TempFilter', 'TempFilterFunctions', 'TempSensor'" echo " Note: 'All' option doesn't include long hual and stress test." echo ' -artifactImageBuildNumber Artifact image build number is used to construct path of docker images, pulling from docker registry. E.g. 20190101.1.' echo " -containerRegistry Host address of container registry." echo " -containerRegistryUsername Username of container registry." echo ' -containerRegistryPassword Password of given username for container registory.' echo ' -iotHubConnectionString IoT hub connection string for creating edge device.' echo ' -eventHubConnectionString Event hub connection string for receive D2C messages.' echo ' -loadGenMessageFrequency Frequency to send messages in LoadGen module for long haul and stress test. Default is 00.00.01 for long haul and 00:00:00.03 for stress test.' echo ' -snitchAlertUrl Alert Url pointing to Azure Logic App for email preparation and sending for long haul and stress test.' echo ' -snitchBuildNumber Build number for snitcher docker image for long haul and stress test. Default is 1.1.' echo ' -snitchReportingIntervalInSecs Reporting frequency in seconds to send status email for long hual and stress test. Default is 86400 (1 day) for long haul and 1700000 for stress test.' echo ' -snitchStorageAccount Azure blob Storage account for store logs used in status email for long haul and stress test.' echo ' -snitchStorageMasterKey Master key of snitch storage account for long haul and stress test.' echo ' -snitchTestDurationInSecs Test duration in seconds for long haul and stress test.' echo ' -loadGen1TransportType Transport type for LoadGen1 for stress test. Default is amqp.' echo ' -loadGen2TransportType Transport type for LoadGen2 for stress test. Default is amqp.' echo ' -loadGen3TransportType Transport type for LoadGen3 for stress test. Default is mqtt.' echo ' -loadGen4TransportType Transport type for LoadGen4 for stress test. Default is mqtt.' echo ' -amqpSettingsEnabled Enable amqp protocol head in Edge Hub.' echo ' -mqttSettingsEnabled Enable mqtt protocol head in Edge Hub.' echo ' -dpsScopeId DPS scope id. Required only when using DPS to provision the device.' echo ' -dpsMasterSymmetricKey DPS master symmetric key. Required only when using DPS symmetric key to provision the Edge device.' echo ' -certScriptDir Optional path to certificate generation script dir' echo ' -installRootCACertPath Optional path to root CA certificate to be used for certificate generation' echo ' -installRootCAKeyPath Optional path to root CA certificate private key to be used for certificate generation' echo ' -installRootCAKeyPassword Optional password to access the root CA certificate private key to be used for certificate generation' echo ' -eventHubConsumerGroupId Optional Event Hub Consumer Group ID for the Analyzer module.' echo ' -desiredModulesToRestartCSV Optional CSV string of module names for long haul specifying what modules to restart. If specified, then "restartIntervalInMins" must be specified as well.' echo ' -restartIntervalInMins Optional value for long haul specifying how often a random module will restart. If specified, then "desiredModulesToRestartCSV" must be specified as well.' echo ' -logAnalyticsEnabled Optional Log Analytics enable string for the Analyzer module. If logAnalyticsEnabled is set to enable (true), the rest of Log Analytics parameters must be provided.' echo ' -logAnalyticsWorkspaceId Optional Log Analytics workspace ID for metrics collection and reporting.' echo ' -logAnalyticsSharedKey Optional Log Analytics shared key for metrics collection and reporting.' echo ' -logAnalyticsLogType Optional Log Analytics log type for the Analyzer module.' exit 1; } process_args "$@" CONTAINER_REGISTRY="${CONTAINER_REGISTRY:-edgebuilds.azurecr.io}" E2E_TEST_DIR="${E2E_TEST_DIR:-$(pwd)}" EVENT_HUB_CONSUMER_GROUP_ID=${EVENT_HUB_CONSUMER_GROUP_ID:-\$Default} SNITCH_BUILD_NUMBER="${SNITCH_BUILD_NUMBER:-1.2}" LOADGEN1_TRANSPORT_TYPE="${LOADGEN1_TRANSPORT_TYPE:-amqp}" LOADGEN2_TRANSPORT_TYPE="${LOADGEN2_TRANSPORT_TYPE:-amqp}" LOADGEN3_TRANSPORT_TYPE="${LOADGEN3_TRANSPORT_TYPE:-mqtt}" LOADGEN4_TRANSPORT_TYPE="${LOADGEN4_TRANSPORT_TYPE:-mqtt}" if [[ "${TEST_NAME,,}" == "longhaul" ]]; then DESIRED_MODULES_TO_RESTART_CSV="${DESIRED_MODULES_TO_RESTART_CSV:-,}" LOADGEN_MESSAGE_FREQUENCY="${LOADGEN_MESSAGE_FREQUENCY:-00:00:01}" RESTART_INTERVAL_IN_MINS="${RESTART_INTERVAL_IN_MINS:-10}" SNITCH_REPORTING_INTERVAL_IN_SECS="${SNITCH_REPORTING_INTERVAL_IN_SECS:-86400}" SNITCH_TEST_DURATION_IN_SECS="${SNITCH_TEST_DURATION_IN_SECS:-604800}" fi if [[ "${TEST_NAME,,}" == "stress" ]]; then LOADGEN_MESSAGE_FREQUENCY="${LOADGEN_MESSAGE_FREQUENCY:-00:00:00.03}" SNITCH_REPORTING_INTERVAL_IN_SECS="${SNITCH_REPORTING_INTERVAL_IN_SECS:-1700000}" SNITCH_TEST_DURATION_IN_SECS="${SNITCH_TEST_DURATION_IN_SECS:-14400}" fi if [ "$AMQP_SETTINGS_ENABLED" != "false" ]; then AMQP_SETTINGS_ENABLED="true" fi if [ "$MQTT_SETTINGS_ENABLED" != "false" ]; then MQTT_SETTINGS_ENABLED="true" fi working_folder="$E2E_TEST_DIR/working" get_image_architecture_label optimize_for_performance=true if [ "$image_architecture_label" = 'arm32v7' ] || [ "$image_architecture_label" = 'arm64v8' ]; then optimize_for_performance=false fi iotedged_artifact_folder="$(get_iotedged_artifact_folder)" iotedge_quickstart_artifact_file="$(get_iotedge_quickstart_artifact_file)" leafdevice_artifact_file="$(get_leafdevice_artifact_file)" twin_testfile_artifact_file="$E2E_TEST_DIR/artifacts/core-linux/e2e_test_files/twin_test_tempSensor.json" module_to_module_deployment_artifact_file="$E2E_TEST_DIR/artifacts/core-linux/e2e_deployment_files/module_to_module_deployment.template.json" module_to_functions_deployment_artifact_file="$E2E_TEST_DIR/artifacts/core-linux/e2e_deployment_files/module_to_functions_deployment.template.json" dm_module_to_module_deployment_artifact_file="$E2E_TEST_DIR/artifacts/core-linux/e2e_deployment_files/dm_module_to_module_deployment.json" long_haul_deployment_artifact_file="$(get_long_haul_deployment_artifact_file)" stress_deployment_artifact_file="$E2E_TEST_DIR/artifacts/core-linux/e2e_deployment_files/stress_deployment.template.json" deployment_working_file="$working_folder/deployment.json" quickstart_working_folder="$working_folder/quickstart" leafdevice_working_folder="$working_folder/leafdevice" run_test
<reponame>112batman/GooseStandalone import { join } from 'path'; import Inquirer from 'inquirer'; import replaceInFile from '../lib/replaceInFile.js'; export default async ({ asarExtractPath }) => { replaceInFile(join(asarExtractPath, 'app_bootstrap', 'Constants.js'), `const UPDATE_ENDPOINT = settings.get('UPDATE_ENDPOINT') || API_ENDPOINT`, `const UPDATE_ENDPOINT = settings.get('UPDATE_ENDPOINT') || 'https://updates.goosemod.com/' + (process.stdout.write('GooseUpdate mods:\\n1. GooseMod\\n2. BetterDiscord\\n3. SmartCord\\n\\nEnter number(s) of mods> ') ? require('child_process').execSync("bash -c 'read -p \\"foobar\\" inp; echo $inp'", {stdio: ['inherit', 'pipe', 'pipe'] }).toString().replace('1', 'goosemod ').replace('2', 'betterdiscord ').replace('3', 'smartcord ').trim().replace(' ', '+') : '')`); replaceInFile(join(asarExtractPath, 'app_bootstrap', 'Constants.js'), `const NEW_UPDATE_ENDPOINT = settings.get('NEW_UPDATE_ENDPOINT') || 'https://discord.com/api/updates/'`, `const NEW_UPDATE_ENDPOINT = settings.get('NEW_UPDATE_ENDPOINT') || (UPDATE_ENDPOINT + '/')`); replaceInFile(join(asarExtractPath, 'common', 'Settings.js'), `return defaultValue`, `return key === 'SKIP_HOST_UPDATE' ? true : defaultValue`); };
<filename>proto/test/v1/proto2/test_all_types/test_all_types.pb.go // Code generated by protoc-gen-go. DO NOT EDIT. // source: proto/test/v1/proto2/test_all_types.proto package test_all_types import ( fmt "fmt" proto "github.com/golang/protobuf/proto" any "github.com/golang/protobuf/ptypes/any" duration "github.com/golang/protobuf/ptypes/duration" _struct "github.com/golang/protobuf/ptypes/struct" timestamp "github.com/golang/protobuf/ptypes/timestamp" wrappers "github.com/golang/protobuf/ptypes/wrappers" math "math" ) // Reference imports to suppress errors if they are not otherwise used. var _ = proto.Marshal var _ = fmt.Errorf var _ = math.Inf // This is a compile-time assertion to ensure that this generated file // is compatible with the proto package it is being compiled against. // A compilation error at this line likely means your copy of the // proto package needs to be updated. const _ = proto.ProtoPackageIsVersion3 // please upgrade the proto package type GlobalEnum int32 const ( GlobalEnum_GOO GlobalEnum = 0 GlobalEnum_GAR GlobalEnum = 1 GlobalEnum_GAZ GlobalEnum = 2 ) var GlobalEnum_name = map[int32]string{ 0: "GOO", 1: "GAR", 2: "GAZ", } var GlobalEnum_value = map[string]int32{ "GOO": 0, "GAR": 1, "GAZ": 2, } func (x GlobalEnum) Enum() *GlobalEnum { p := new(GlobalEnum) *p = x return p } func (x GlobalEnum) String() string { return proto.EnumName(GlobalEnum_name, int32(x)) } func (x *GlobalEnum) UnmarshalJSON(data []byte) error { value, err := proto.UnmarshalJSONEnum(GlobalEnum_value, data, "GlobalEnum") if err != nil { return err } *x = GlobalEnum(value) return nil } func (GlobalEnum) EnumDescriptor() ([]byte, []int) { return fileDescriptor_98cc99a98b5d98fe, []int{0} } type TestAllTypes_NestedEnum int32 const ( TestAllTypes_FOO TestAllTypes_NestedEnum = 0 TestAllTypes_BAR TestAllTypes_NestedEnum = 1 TestAllTypes_BAZ TestAllTypes_NestedEnum = 2 ) var TestAllTypes_NestedEnum_name = map[int32]string{ 0: "FOO", 1: "BAR", 2: "BAZ", } var TestAllTypes_NestedEnum_value = map[string]int32{ "FOO": 0, "BAR": 1, "BAZ": 2, } func (x TestAllTypes_NestedEnum) Enum() *TestAllTypes_NestedEnum { p := new(TestAllTypes_NestedEnum) *p = x return p } func (x TestAllTypes_NestedEnum) String() string { return proto.EnumName(TestAllTypes_NestedEnum_name, int32(x)) } func (x *TestAllTypes_NestedEnum) UnmarshalJSON(data []byte) error { value, err := proto.UnmarshalJSONEnum(TestAllTypes_NestedEnum_value, data, "TestAllTypes_NestedEnum") if err != nil { return err } *x = TestAllTypes_NestedEnum(value) return nil } func (TestAllTypes_NestedEnum) EnumDescriptor() ([]byte, []int) { return fileDescriptor_98cc99a98b5d98fe, []int{0, 0} } type TestAllTypes struct { SingleInt32 *int32 `protobuf:"varint,1,opt,name=single_int32,json=singleInt32,def=-32" json:"single_int32,omitempty"` SingleInt64 *int64 `protobuf:"varint,2,opt,name=single_int64,json=singleInt64,def=-64" json:"single_int64,omitempty"` SingleUint32 *uint32 `protobuf:"varint,3,opt,name=single_uint32,json=singleUint32,def=32" json:"single_uint32,omitempty"` SingleUint64 *uint64 `protobuf:"varint,4,opt,name=single_uint64,json=singleUint64,def=64" json:"single_uint64,omitempty"` SingleSint32 *int32 `protobuf:"zigzag32,5,opt,name=single_sint32,json=singleSint32" json:"single_sint32,omitempty"` SingleSint64 *int64 `protobuf:"zigzag64,6,opt,name=single_sint64,json=singleSint64" json:"single_sint64,omitempty"` SingleFixed32 *uint32 `protobuf:"fixed32,7,opt,name=single_fixed32,json=singleFixed32" json:"single_fixed32,omitempty"` SingleFixed64 *uint64 `protobuf:"fixed64,8,opt,name=single_fixed64,json=singleFixed64" json:"single_fixed64,omitempty"` SingleSfixed32 *int32 `protobuf:"fixed32,9,opt,name=single_sfixed32,json=singleSfixed32" json:"single_sfixed32,omitempty"` SingleSfixed64 *int64 `protobuf:"fixed64,10,opt,name=single_sfixed64,json=singleSfixed64" json:"single_sfixed64,omitempty"` SingleFloat *float32 `protobuf:"fixed32,11,opt,name=single_float,json=singleFloat,def=3" json:"single_float,omitempty"` SingleDouble *float64 `protobuf:"fixed64,12,opt,name=single_double,json=singleDouble,def=6.4" json:"single_double,omitempty"` SingleBool *bool `protobuf:"varint,13,opt,name=single_bool,json=singleBool,def=1" json:"single_bool,omitempty"` SingleString *string `protobuf:"bytes,14,opt,name=single_string,json=singleString,def=empty" json:"single_string,omitempty"` SingleBytes []byte `protobuf:"bytes,15,opt,name=single_bytes,json=singleBytes,def=none" json:"single_bytes,omitempty"` SingleAny *any.Any `protobuf:"bytes,100,opt,name=single_any,json=singleAny" json:"single_any,omitempty"` SingleDuration *duration.Duration `protobuf:"bytes,101,opt,name=single_duration,json=singleDuration" json:"single_duration,omitempty"` SingleTimestamp *timestamp.Timestamp `protobuf:"bytes,102,opt,name=single_timestamp,json=singleTimestamp" json:"single_timestamp,omitempty"` SingleStruct *_struct.Struct `protobuf:"bytes,103,opt,name=single_struct,json=singleStruct" json:"single_struct,omitempty"` SingleValue *_struct.Value `protobuf:"bytes,104,opt,name=single_value,json=singleValue" json:"single_value,omitempty"` SingleInt64Wrapper *wrappers.Int64Value `protobuf:"bytes,105,opt,name=single_int64_wrapper,json=singleInt64Wrapper" json:"single_int64_wrapper,omitempty"` SingleInt32Wrapper *wrappers.Int32Value `protobuf:"bytes,106,opt,name=single_int32_wrapper,json=singleInt32Wrapper" json:"single_int32_wrapper,omitempty"` SingleDoubleWrapper *wrappers.DoubleValue `protobuf:"bytes,107,opt,name=single_double_wrapper,json=singleDoubleWrapper" json:"single_double_wrapper,omitempty"` SingleFloatWrapper *wrappers.FloatValue `protobuf:"bytes,108,opt,name=single_float_wrapper,json=singleFloatWrapper" json:"single_float_wrapper,omitempty"` SingleUint64Wrapper *wrappers.UInt64Value `protobuf:"bytes,109,opt,name=single_uint64_wrapper,json=singleUint64Wrapper" json:"single_uint64_wrapper,omitempty"` SingleUint32Wrapper *wrappers.UInt32Value `protobuf:"bytes,110,opt,name=single_uint32_wrapper,json=singleUint32Wrapper" json:"single_uint32_wrapper,omitempty"` SingleStringWrapper *wrappers.StringValue `protobuf:"bytes,111,opt,name=single_string_wrapper,json=singleStringWrapper" json:"single_string_wrapper,omitempty"` SingleBoolWrapper *wrappers.BoolValue `protobuf:"bytes,112,opt,name=single_bool_wrapper,json=singleBoolWrapper" json:"single_bool_wrapper,omitempty"` SingleBytesWrapper *wrappers.BytesValue `protobuf:"bytes,113,opt,name=single_bytes_wrapper,json=singleBytesWrapper" json:"single_bytes_wrapper,omitempty"` // Types that are valid to be assigned to NestedType: // *TestAllTypes_SingleNestedMessage // *TestAllTypes_SingleNestedEnum NestedType isTestAllTypes_NestedType `protobuf_oneof:"nested_type"` StandaloneMessage *TestAllTypes_NestedMessage `protobuf:"bytes,23,opt,name=standalone_message,json=standaloneMessage" json:"standalone_message,omitempty"` StandaloneEnum *TestAllTypes_NestedEnum `protobuf:"varint,24,opt,name=standalone_enum,json=standaloneEnum,enum=google.api.expr.test.v1.proto2.TestAllTypes_NestedEnum" json:"standalone_enum,omitempty"` RepeatedInt32 []int32 `protobuf:"varint,31,rep,name=repeated_int32,json=repeatedInt32" json:"repeated_int32,omitempty"` RepeatedInt64 []int64 `protobuf:"varint,32,rep,name=repeated_int64,json=repeatedInt64" json:"repeated_int64,omitempty"` RepeatedUint32 []uint32 `protobuf:"varint,33,rep,name=repeated_uint32,json=repeatedUint32" json:"repeated_uint32,omitempty"` RepeatedUint64 []uint64 `protobuf:"varint,34,rep,name=repeated_uint64,json=repeatedUint64" json:"repeated_uint64,omitempty"` RepeatedSint32 []int32 `protobuf:"zigzag32,35,rep,name=repeated_sint32,json=repeatedSint32" json:"repeated_sint32,omitempty"` RepeatedSint64 []int64 `protobuf:"zigzag64,36,rep,name=repeated_sint64,json=repeatedSint64" json:"repeated_sint64,omitempty"` RepeatedFixed32 []uint32 `protobuf:"fixed32,37,rep,name=repeated_fixed32,json=repeatedFixed32" json:"repeated_fixed32,omitempty"` RepeatedFixed64 []uint64 `protobuf:"fixed64,38,rep,name=repeated_fixed64,json=repeatedFixed64" json:"repeated_fixed64,omitempty"` RepeatedSfixed32 []int32 `protobuf:"fixed32,39,rep,name=repeated_sfixed32,json=repeatedSfixed32" json:"repeated_sfixed32,omitempty"` RepeatedSfixed64 []int64 `protobuf:"fixed64,40,rep,name=repeated_sfixed64,json=repeatedSfixed64" json:"repeated_sfixed64,omitempty"` RepeatedFloat []float32 `protobuf:"fixed32,41,rep,name=repeated_float,json=repeatedFloat" json:"repeated_float,omitempty"` RepeatedDouble []float64 `protobuf:"fixed64,42,rep,name=repeated_double,json=repeatedDouble" json:"repeated_double,omitempty"` RepeatedBool []bool `protobuf:"varint,43,rep,name=repeated_bool,json=repeatedBool" json:"repeated_bool,omitempty"` RepeatedString []string `protobuf:"bytes,44,rep,name=repeated_string,json=repeatedString" json:"repeated_string,omitempty"` RepeatedBytes [][]byte `protobuf:"bytes,45,rep,name=repeated_bytes,json=repeatedBytes" json:"repeated_bytes,omitempty"` RepeatedNestedMessage []*TestAllTypes_NestedMessage `protobuf:"bytes,51,rep,name=repeated_nested_message,json=repeatedNestedMessage" json:"repeated_nested_message,omitempty"` RepeatedNestedEnum []TestAllTypes_NestedEnum `protobuf:"varint,52,rep,name=repeated_nested_enum,json=repeatedNestedEnum,enum=google.api.expr.test.v1.proto2.TestAllTypes_NestedEnum" json:"repeated_nested_enum,omitempty"` RepeatedStringPiece []string `protobuf:"bytes,53,rep,name=repeated_string_piece,json=repeatedStringPiece" json:"repeated_string_piece,omitempty"` RepeatedCord []string `protobuf:"bytes,54,rep,name=repeated_cord,json=repeatedCord" json:"repeated_cord,omitempty"` RepeatedLazyMessage []*TestAllTypes_NestedMessage `protobuf:"bytes,55,rep,name=repeated_lazy_message,json=repeatedLazyMessage" json:"repeated_lazy_message,omitempty"` MapStringString map[string]string `protobuf:"bytes,61,rep,name=map_string_string,json=mapStringString" json:"map_string_string,omitempty" protobuf_key:"bytes,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` MapInt64NestedType map[int64]*NestedTestAllTypes `protobuf:"bytes,62,rep,name=map_int64_nested_type,json=mapInt64NestedType" json:"map_int64_nested_type,omitempty" protobuf_key:"varint,1,opt,name=key" protobuf_val:"bytes,2,opt,name=value"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *TestAllTypes) Reset() { *m = TestAllTypes{} } func (m *TestAllTypes) String() string { return proto.CompactTextString(m) } func (*TestAllTypes) ProtoMessage() {} func (*TestAllTypes) Descriptor() ([]byte, []int) { return fileDescriptor_98cc99a98b5d98fe, []int{0} } func (m *TestAllTypes) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TestAllTypes.Unmarshal(m, b) } func (m *TestAllTypes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_TestAllTypes.Marshal(b, m, deterministic) } func (m *TestAllTypes) XXX_Merge(src proto.Message) { xxx_messageInfo_TestAllTypes.Merge(m, src) } func (m *TestAllTypes) XXX_Size() int { return xxx_messageInfo_TestAllTypes.Size(m) } func (m *TestAllTypes) XXX_DiscardUnknown() { xxx_messageInfo_TestAllTypes.DiscardUnknown(m) } var xxx_messageInfo_TestAllTypes proto.InternalMessageInfo const Default_TestAllTypes_SingleInt32 int32 = -32 const Default_TestAllTypes_SingleInt64 int64 = -64 const Default_TestAllTypes_SingleUint32 uint32 = 32 const Default_TestAllTypes_SingleUint64 uint64 = 64 const Default_TestAllTypes_SingleFloat float32 = 3 const Default_TestAllTypes_SingleDouble float64 = 6.4 const Default_TestAllTypes_SingleBool bool = true const Default_TestAllTypes_SingleString string = "empty" var Default_TestAllTypes_SingleBytes []byte = []byte("none") const Default_TestAllTypes_SingleNestedEnum TestAllTypes_NestedEnum = TestAllTypes_BAR func (m *TestAllTypes) GetSingleInt32() int32 { if m != nil && m.SingleInt32 != nil { return *m.SingleInt32 } return Default_TestAllTypes_SingleInt32 } func (m *TestAllTypes) GetSingleInt64() int64 { if m != nil && m.SingleInt64 != nil { return *m.SingleInt64 } return Default_TestAllTypes_SingleInt64 } func (m *TestAllTypes) GetSingleUint32() uint32 { if m != nil && m.SingleUint32 != nil { return *m.SingleUint32 } return Default_TestAllTypes_SingleUint32 } func (m *TestAllTypes) GetSingleUint64() uint64 { if m != nil && m.SingleUint64 != nil { return *m.SingleUint64 } return Default_TestAllTypes_SingleUint64 } func (m *TestAllTypes) GetSingleSint32() int32 { if m != nil && m.SingleSint32 != nil { return *m.SingleSint32 } return 0 } func (m *TestAllTypes) GetSingleSint64() int64 { if m != nil && m.SingleSint64 != nil { return *m.SingleSint64 } return 0 } func (m *TestAllTypes) GetSingleFixed32() uint32 { if m != nil && m.SingleFixed32 != nil { return *m.SingleFixed32 } return 0 } func (m *TestAllTypes) GetSingleFixed64() uint64 { if m != nil && m.SingleFixed64 != nil { return *m.SingleFixed64 } return 0 } func (m *TestAllTypes) GetSingleSfixed32() int32 { if m != nil && m.SingleSfixed32 != nil { return *m.SingleSfixed32 } return 0 } func (m *TestAllTypes) GetSingleSfixed64() int64 { if m != nil && m.SingleSfixed64 != nil { return *m.SingleSfixed64 } return 0 } func (m *TestAllTypes) GetSingleFloat() float32 { if m != nil && m.SingleFloat != nil { return *m.SingleFloat } return Default_TestAllTypes_SingleFloat } func (m *TestAllTypes) GetSingleDouble() float64 { if m != nil && m.SingleDouble != nil { return *m.SingleDouble } return Default_TestAllTypes_SingleDouble } func (m *TestAllTypes) GetSingleBool() bool { if m != nil && m.SingleBool != nil { return *m.SingleBool } return Default_TestAllTypes_SingleBool } func (m *TestAllTypes) GetSingleString() string { if m != nil && m.SingleString != nil { return *m.SingleString } return Default_TestAllTypes_SingleString } func (m *TestAllTypes) GetSingleBytes() []byte { if m != nil && m.SingleBytes != nil { return m.SingleBytes } return append([]byte(nil), Default_TestAllTypes_SingleBytes...) } func (m *TestAllTypes) GetSingleAny() *any.Any { if m != nil { return m.SingleAny } return nil } func (m *TestAllTypes) GetSingleDuration() *duration.Duration { if m != nil { return m.SingleDuration } return nil } func (m *TestAllTypes) GetSingleTimestamp() *timestamp.Timestamp { if m != nil { return m.SingleTimestamp } return nil } func (m *TestAllTypes) GetSingleStruct() *_struct.Struct { if m != nil { return m.SingleStruct } return nil } func (m *TestAllTypes) GetSingleValue() *_struct.Value { if m != nil { return m.SingleValue } return nil } func (m *TestAllTypes) GetSingleInt64Wrapper() *wrappers.Int64Value { if m != nil { return m.SingleInt64Wrapper } return nil } func (m *TestAllTypes) GetSingleInt32Wrapper() *wrappers.Int32Value { if m != nil { return m.SingleInt32Wrapper } return nil } func (m *TestAllTypes) GetSingleDoubleWrapper() *wrappers.DoubleValue { if m != nil { return m.SingleDoubleWrapper } return nil } func (m *TestAllTypes) GetSingleFloatWrapper() *wrappers.FloatValue { if m != nil { return m.SingleFloatWrapper } return nil } func (m *TestAllTypes) GetSingleUint64Wrapper() *wrappers.UInt64Value { if m != nil { return m.SingleUint64Wrapper } return nil } func (m *TestAllTypes) GetSingleUint32Wrapper() *wrappers.UInt32Value { if m != nil { return m.SingleUint32Wrapper } return nil } func (m *TestAllTypes) GetSingleStringWrapper() *wrappers.StringValue { if m != nil { return m.SingleStringWrapper } return nil } func (m *TestAllTypes) GetSingleBoolWrapper() *wrappers.BoolValue { if m != nil { return m.SingleBoolWrapper } return nil } func (m *TestAllTypes) GetSingleBytesWrapper() *wrappers.BytesValue { if m != nil { return m.SingleBytesWrapper } return nil } type isTestAllTypes_NestedType interface { isTestAllTypes_NestedType() } type TestAllTypes_SingleNestedMessage struct { SingleNestedMessage *TestAllTypes_NestedMessage `protobuf:"bytes,21,opt,name=single_nested_message,json=singleNestedMessage,oneof"` } type TestAllTypes_SingleNestedEnum struct { SingleNestedEnum TestAllTypes_NestedEnum `protobuf:"varint,22,opt,name=single_nested_enum,json=singleNestedEnum,enum=google.api.expr.test.v1.proto2.TestAllTypes_NestedEnum,oneof,def=1"` } func (*TestAllTypes_SingleNestedMessage) isTestAllTypes_NestedType() {} func (*TestAllTypes_SingleNestedEnum) isTestAllTypes_NestedType() {} func (m *TestAllTypes) GetNestedType() isTestAllTypes_NestedType { if m != nil { return m.NestedType } return nil } func (m *TestAllTypes) GetSingleNestedMessage() *TestAllTypes_NestedMessage { if x, ok := m.GetNestedType().(*TestAllTypes_SingleNestedMessage); ok { return x.SingleNestedMessage } return nil } func (m *TestAllTypes) GetSingleNestedEnum() TestAllTypes_NestedEnum { if x, ok := m.GetNestedType().(*TestAllTypes_SingleNestedEnum); ok { return x.SingleNestedEnum } return Default_TestAllTypes_SingleNestedEnum } func (m *TestAllTypes) GetStandaloneMessage() *TestAllTypes_NestedMessage { if m != nil { return m.StandaloneMessage } return nil } func (m *TestAllTypes) GetStandaloneEnum() TestAllTypes_NestedEnum { if m != nil && m.StandaloneEnum != nil { return *m.StandaloneEnum } return TestAllTypes_FOO } func (m *TestAllTypes) GetRepeatedInt32() []int32 { if m != nil { return m.RepeatedInt32 } return nil } func (m *TestAllTypes) GetRepeatedInt64() []int64 { if m != nil { return m.RepeatedInt64 } return nil } func (m *TestAllTypes) GetRepeatedUint32() []uint32 { if m != nil { return m.RepeatedUint32 } return nil } func (m *TestAllTypes) GetRepeatedUint64() []uint64 { if m != nil { return m.RepeatedUint64 } return nil } func (m *TestAllTypes) GetRepeatedSint32() []int32 { if m != nil { return m.RepeatedSint32 } return nil } func (m *TestAllTypes) GetRepeatedSint64() []int64 { if m != nil { return m.RepeatedSint64 } return nil } func (m *TestAllTypes) GetRepeatedFixed32() []uint32 { if m != nil { return m.RepeatedFixed32 } return nil } func (m *TestAllTypes) GetRepeatedFixed64() []uint64 { if m != nil { return m.RepeatedFixed64 } return nil } func (m *TestAllTypes) GetRepeatedSfixed32() []int32 { if m != nil { return m.RepeatedSfixed32 } return nil } func (m *TestAllTypes) GetRepeatedSfixed64() []int64 { if m != nil { return m.RepeatedSfixed64 } return nil } func (m *TestAllTypes) GetRepeatedFloat() []float32 { if m != nil { return m.RepeatedFloat } return nil } func (m *TestAllTypes) GetRepeatedDouble() []float64 { if m != nil { return m.RepeatedDouble } return nil } func (m *TestAllTypes) GetRepeatedBool() []bool { if m != nil { return m.RepeatedBool } return nil } func (m *TestAllTypes) GetRepeatedString() []string { if m != nil { return m.RepeatedString } return nil } func (m *TestAllTypes) GetRepeatedBytes() [][]byte { if m != nil { return m.RepeatedBytes } return nil } func (m *TestAllTypes) GetRepeatedNestedMessage() []*TestAllTypes_NestedMessage { if m != nil { return m.RepeatedNestedMessage } return nil } func (m *TestAllTypes) GetRepeatedNestedEnum() []TestAllTypes_NestedEnum { if m != nil { return m.RepeatedNestedEnum } return nil } func (m *TestAllTypes) GetRepeatedStringPiece() []string { if m != nil { return m.RepeatedStringPiece } return nil } func (m *TestAllTypes) GetRepeatedCord() []string { if m != nil { return m.RepeatedCord } return nil } func (m *TestAllTypes) GetRepeatedLazyMessage() []*TestAllTypes_NestedMessage { if m != nil { return m.RepeatedLazyMessage } return nil } func (m *TestAllTypes) GetMapStringString() map[string]string { if m != nil { return m.MapStringString } return nil } func (m *TestAllTypes) GetMapInt64NestedType() map[int64]*NestedTestAllTypes { if m != nil { return m.MapInt64NestedType } return nil } // XXX_OneofWrappers is for the internal use of the proto package. func (*TestAllTypes) XXX_OneofWrappers() []interface{} { return []interface{}{ (*TestAllTypes_SingleNestedMessage)(nil), (*TestAllTypes_SingleNestedEnum)(nil), } } type TestAllTypes_NestedMessage struct { Bb *int32 `protobuf:"varint,1,opt,name=bb" json:"bb,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *TestAllTypes_NestedMessage) Reset() { *m = TestAllTypes_NestedMessage{} } func (m *TestAllTypes_NestedMessage) String() string { return proto.CompactTextString(m) } func (*TestAllTypes_NestedMessage) ProtoMessage() {} func (*TestAllTypes_NestedMessage) Descriptor() ([]byte, []int) { return fileDescriptor_98cc99a98b5d98fe, []int{0, 0} } func (m *TestAllTypes_NestedMessage) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TestAllTypes_NestedMessage.Unmarshal(m, b) } func (m *TestAllTypes_NestedMessage) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_TestAllTypes_NestedMessage.Marshal(b, m, deterministic) } func (m *TestAllTypes_NestedMessage) XXX_Merge(src proto.Message) { xxx_messageInfo_TestAllTypes_NestedMessage.Merge(m, src) } func (m *TestAllTypes_NestedMessage) XXX_Size() int { return xxx_messageInfo_TestAllTypes_NestedMessage.Size(m) } func (m *TestAllTypes_NestedMessage) XXX_DiscardUnknown() { xxx_messageInfo_TestAllTypes_NestedMessage.DiscardUnknown(m) } var xxx_messageInfo_TestAllTypes_NestedMessage proto.InternalMessageInfo func (m *TestAllTypes_NestedMessage) GetBb() int32 { if m != nil && m.Bb != nil { return *m.Bb } return 0 } type NestedTestAllTypes struct { Child *NestedTestAllTypes `protobuf:"bytes,1,opt,name=child" json:"child,omitempty"` Payload *TestAllTypes `protobuf:"bytes,2,opt,name=payload" json:"payload,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *NestedTestAllTypes) Reset() { *m = NestedTestAllTypes{} } func (m *NestedTestAllTypes) String() string { return proto.CompactTextString(m) } func (*NestedTestAllTypes) ProtoMessage() {} func (*NestedTestAllTypes) Descriptor() ([]byte, []int) { return fileDescriptor_98cc99a98b5d98fe, []int{1} } func (m *NestedTestAllTypes) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_NestedTestAllTypes.Unmarshal(m, b) } func (m *NestedTestAllTypes) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_NestedTestAllTypes.Marshal(b, m, deterministic) } func (m *NestedTestAllTypes) XXX_Merge(src proto.Message) { xxx_messageInfo_NestedTestAllTypes.Merge(m, src) } func (m *NestedTestAllTypes) XXX_Size() int { return xxx_messageInfo_NestedTestAllTypes.Size(m) } func (m *NestedTestAllTypes) XXX_DiscardUnknown() { xxx_messageInfo_NestedTestAllTypes.DiscardUnknown(m) } var xxx_messageInfo_NestedTestAllTypes proto.InternalMessageInfo func (m *NestedTestAllTypes) GetChild() *NestedTestAllTypes { if m != nil { return m.Child } return nil } func (m *NestedTestAllTypes) GetPayload() *TestAllTypes { if m != nil { return m.Payload } return nil } type TestRequired struct { RequiredInt32 *int32 `protobuf:"varint,1,req,name=required_int32,json=requiredInt32" json:"required_int32,omitempty"` XXX_NoUnkeyedLiteral struct{} `json:"-"` XXX_unrecognized []byte `json:"-"` XXX_sizecache int32 `json:"-"` } func (m *TestRequired) Reset() { *m = TestRequired{} } func (m *TestRequired) String() string { return proto.CompactTextString(m) } func (*TestRequired) ProtoMessage() {} func (*TestRequired) Descriptor() ([]byte, []int) { return fileDescriptor_98cc99a98b5d98fe, []int{2} } func (m *TestRequired) XXX_Unmarshal(b []byte) error { return xxx_messageInfo_TestRequired.Unmarshal(m, b) } func (m *TestRequired) XXX_Marshal(b []byte, deterministic bool) ([]byte, error) { return xxx_messageInfo_TestRequired.Marshal(b, m, deterministic) } func (m *TestRequired) XXX_Merge(src proto.Message) { xxx_messageInfo_TestRequired.Merge(m, src) } func (m *TestRequired) XXX_Size() int { return xxx_messageInfo_TestRequired.Size(m) } func (m *TestRequired) XXX_DiscardUnknown() { xxx_messageInfo_TestRequired.DiscardUnknown(m) } var xxx_messageInfo_TestRequired proto.InternalMessageInfo func (m *TestRequired) GetRequiredInt32() int32 { if m != nil && m.RequiredInt32 != nil { return *m.RequiredInt32 } return 0 } func init() { proto.RegisterEnum("google.api.expr.test.v1.proto2.GlobalEnum", GlobalEnum_name, GlobalEnum_value) proto.RegisterEnum("google.api.expr.test.v1.proto2.TestAllTypes_NestedEnum", TestAllTypes_NestedEnum_name, TestAllTypes_NestedEnum_value) proto.RegisterType((*TestAllTypes)(nil), "google.api.expr.test.v1.proto2.TestAllTypes") proto.RegisterMapType((map[int64]*NestedTestAllTypes)(nil), "google.api.expr.test.v1.proto2.TestAllTypes.MapInt64NestedTypeEntry") proto.RegisterMapType((map[string]string)(nil), "google.api.expr.test.v1.proto2.TestAllTypes.MapStringStringEntry") proto.RegisterType((*TestAllTypes_NestedMessage)(nil), "google.api.expr.test.v1.proto2.TestAllTypes.NestedMessage") proto.RegisterType((*NestedTestAllTypes)(nil), "google.api.expr.test.v1.proto2.NestedTestAllTypes") proto.RegisterType((*TestRequired)(nil), "google.api.expr.test.v1.proto2.TestRequired") } func init() { proto.RegisterFile("proto/test/v1/proto2/test_all_types.proto", fileDescriptor_98cc99a98b5d98fe) } var fileDescriptor_98cc99a98b5d98fe = []byte{ // 1403 bytes of a gzipped FileDescriptorProto 0x1f, 0x8b, 0x08, 0x00, 0x00, 0x00, 0x00, 0x00, 0x02, 0xff, 0xa4, 0x58, 0x7d, 0x73, 0xd3, 0xc6, 0x13, 0x46, 0xba, 0x84, 0x24, 0x97, 0x38, 0x71, 0x2e, 0x09, 0x39, 0xf2, 0x63, 0xe0, 0x7e, 0x81, 0x90, 0xe3, 0x25, 0xce, 0x60, 0x0b, 0x41, 0x3d, 0x6d, 0x67, 0xe2, 0x42, 0x42, 0x3b, 0xa5, 0x30, 0x0a, 0xb4, 0x33, 0xfc, 0xe3, 0xca, 0xf6, 0xc5, 0xa8, 0xc8, 0x92, 0x90, 0x25, 0xc0, 0x7c, 0xa0, 0x7e, 0x85, 0x7e, 0xbd, 0x8e, 0xee, 0xcd, 0x7a, 0xc1, 0x74, 0x42, 0xfe, 0xc9, 0x58, 0xbb, 0xcf, 0x3d, 0xbb, 0xfb, 0xe8, 0xf6, 0x6e, 0x15, 0x78, 0x27, 0x8a, 0xc3, 0x24, 0x3c, 0x4c, 0xd8, 0x38, 0x39, 0xfc, 0xf0, 0xe0, 0x90, 0x3f, 0x35, 0xf9, 0x63, 0xd7, 0xf5, 0xfd, 0x6e, 0x32, 0x89, 0xd8, 0xb8, 0xc1, 0xad, 0xe8, 0xfa, 0x30, 0x0c, 0x87, 0x3e, 0x6b, 0xb8, 0x91, 0xd7, 0x60, 0x9f, 0xa2, 0xb8, 0x91, 0xa1, 0x1a, 0x1f, 0x1e, 0x08, 0x77, 0x73, 0xe7, 0xaa, 0xf0, 0x0b, 0x8e, 0x5e, 0x7a, 0x76, 0xe8, 0x06, 0x13, 0xe1, 0xdb, 0xb9, 0x5e, 0x76, 0x0d, 0xd2, 0xd8, 0x4d, 0xbc, 0x30, 0x90, 0xfe, 0x6b, 0x65, 0xff, 0x38, 0x89, 0xd3, 0x7e, 0x22, 0xbd, 0x37, 0xca, 0xde, 0xc4, 0x1b, 0xb1, 0x71, 0xe2, 0x8e, 0xa2, 0x59, 0xf4, 0x1f, 0x63, 0x37, 0x8a, 0x58, 0x2c, 0x33, 0xdf, 0xfd, 0xe7, 0x1a, 0x5c, 0x79, 0xc5, 0xc6, 0xc9, 0x91, 0xef, 0xbf, 0xca, 0x0a, 0x42, 0xb7, 0xe1, 0xca, 0xd8, 0x0b, 0x86, 0x3e, 0xeb, 0x7a, 0x41, 0xd2, 0x6a, 0x62, 0x83, 0x18, 0x74, 0xbe, 0x0d, 0x0e, 0x5a, 0x4d, 0x67, 0x59, 0x38, 0x7e, 0xce, 0xec, 0x45, 0x9c, 0x6d, 0x61, 0x93, 0x18, 0x14, 0xb4, 0xc1, 0x81, 0x6d, 0xe5, 0x70, 0xb6, 0x85, 0xf6, 0x61, 0x4d, 0xe2, 0x52, 0x41, 0x08, 0x88, 0x41, 0x6b, 0x6d, 0xb3, 0xd5, 0x74, 0x24, 0xc1, 0x6b, 0x6e, 0x2f, 0x01, 0x6d, 0x0b, 0xcf, 0x11, 0x83, 0xce, 0xb5, 0x4d, 0xdb, 0xca, 0x03, 0x6d, 0x0b, 0xdd, 0xd4, 0xc0, 0xb1, 0x60, 0x9c, 0x27, 0x06, 0x5d, 0x57, 0xa0, 0x53, 0xc1, 0x56, 0x04, 0xd9, 0x16, 0xbe, 0x4c, 0x0c, 0x8a, 0xf2, 0x20, 0xdb, 0x42, 0x7b, 0x70, 0x55, 0x82, 0xce, 0xbc, 0x4f, 0x6c, 0xd0, 0x6a, 0xe2, 0x05, 0x62, 0xd0, 0x05, 0x47, 0x2e, 0x3d, 0x16, 0xc6, 0x32, 0xcc, 0xb6, 0xf0, 0x22, 0x31, 0xe8, 0xe5, 0x02, 0x8c, 0x57, 0xba, 0xa6, 0x42, 0x2a, 0xba, 0x25, 0x62, 0xd0, 0x35, 0x47, 0xae, 0x3e, 0x95, 0xd6, 0x0a, 0xd0, 0xb6, 0x30, 0x24, 0x06, 0xad, 0x17, 0x81, 0xb6, 0x85, 0x6e, 0x69, 0x8d, 0xcf, 0xfc, 0xd0, 0x4d, 0xf0, 0x32, 0x31, 0xa8, 0xd9, 0x36, 0x5a, 0x4a, 0xe1, 0xe3, 0xcc, 0x8a, 0xa8, 0x2e, 0x75, 0x10, 0xa6, 0x3d, 0x9f, 0xe1, 0x15, 0x62, 0x50, 0xa3, 0x0d, 0xec, 0x86, 0x56, 0xee, 0x09, 0x77, 0xa0, 0x3d, 0x28, 0x17, 0x76, 0x7b, 0x61, 0xe8, 0xe3, 0x1a, 0x31, 0xe8, 0x62, 0x7b, 0x2e, 0x89, 0x53, 0xe6, 0x40, 0xe1, 0xe8, 0x84, 0xa1, 0x8f, 0xee, 0x4e, 0xb5, 0x4b, 0x62, 0x2f, 0x18, 0xe2, 0x55, 0x62, 0xd0, 0xa5, 0xf6, 0x3c, 0x1b, 0x45, 0xc9, 0x44, 0x4b, 0xc8, 0x5d, 0x68, 0x5f, 0xa7, 0xd8, 0x9b, 0x24, 0x6c, 0x8c, 0xd7, 0x88, 0x41, 0x57, 0xda, 0x73, 0x41, 0x18, 0x30, 0x95, 0x65, 0x27, 0x73, 0xa0, 0x16, 0x94, 0x21, 0xba, 0x6e, 0x30, 0xc1, 0x03, 0x62, 0xd0, 0xe5, 0xe6, 0x66, 0x43, 0xf6, 0x8d, 0xda, 0x9d, 0x8d, 0xa3, 0x60, 0xe2, 0x2c, 0x09, 0xdc, 0x51, 0x30, 0x41, 0x1d, 0xad, 0x94, 0xea, 0x0a, 0xcc, 0xf8, 0xca, 0xab, 0x95, 0x95, 0x4f, 0x24, 0x40, 0x89, 0xa8, 0x9e, 0xd1, 0x53, 0x58, 0x97, 0x1c, 0xba, 0x37, 0xf0, 0x19, 0x27, 0xd9, 0xa9, 0x90, 0xbc, 0x52, 0x08, 0x47, 0xc6, 0xd5, 0x06, 0xf4, 0x7d, 0x5e, 0x94, 0xb4, 0x9f, 0xe0, 0x21, 0xe7, 0xd8, 0xae, 0x70, 0x9c, 0x72, 0x77, 0x4e, 0xa6, 0xb4, 0x9f, 0xa0, 0xef, 0xb4, 0x4c, 0x1f, 0x5c, 0x3f, 0x65, 0xf8, 0x2d, 0x5f, 0x7c, 0xa5, 0xb2, 0xf8, 0xf7, 0xcc, 0xab, 0x84, 0xe3, 0x0f, 0xe8, 0x39, 0xdc, 0xcc, 0x37, 0x5a, 0x57, 0x36, 0x30, 0xf6, 0x38, 0xc5, 0xff, 0x2a, 0x14, 0xbc, 0xed, 0x04, 0x0f, 0xca, 0x35, 0xe2, 0x1f, 0x62, 0x59, 0x91, 0xae, 0xd5, 0xd4, 0x74, 0x7f, 0xcd, 0xa6, 0x6b, 0x35, 0xcb, 0x74, 0xad, 0xa6, 0xa2, 0x7b, 0x09, 0xb7, 0x0a, 0x9b, 0x4f, 0xf3, 0xbd, 0xe3, 0x7c, 0xd7, 0xaa, 0xef, 0x89, 0xc3, 0x04, 0xe1, 0x46, 0x7e, 0x77, 0x56, 0x13, 0xe4, 0x9b, 0x5e, 0x13, 0xfa, 0x33, 0x12, 0xe4, 0x4d, 0x50, 0x48, 0x90, 0x5b, 0xaa, 0x09, 0xa6, 0x45, 0xfd, 0x46, 0x33, 0x12, 0x7c, 0x9d, 0x13, 0x70, 0x23, 0x7f, 0xf0, 0x7c, 0x99, 0x31, 0x27, 0x61, 0xf0, 0x15, 0x46, 0xa5, 0xe1, 0x46, 0xfe, 0xcc, 0xab, 0x32, 0x8a, 0x86, 0xd3, 0x8c, 0xe1, 0x0c, 0x46, 0xd1, 0x7c, 0x05, 0x46, 0x61, 0x52, 0x8c, 0xbf, 0xc0, 0x8d, 0x5c, 0xa7, 0x6b, 0xbe, 0x68, 0xc6, 0xbe, 0xcf, 0xda, 0x5e, 0xb0, 0xad, 0x4f, 0xcf, 0x81, 0xea, 0x0b, 0xe1, 0x2d, 0xae, 0xc9, 0xde, 0xcf, 0x78, 0x21, 0xbc, 0xdf, 0x0b, 0x2f, 0x84, 0x5b, 0x14, 0x5d, 0xa4, 0x8b, 0x0d, 0xd8, 0x38, 0x61, 0x83, 0xee, 0x88, 0x8d, 0xc7, 0xee, 0x90, 0xe1, 0x2d, 0xce, 0xd7, 0x6e, 0x7c, 0xfd, 0x2e, 0x6d, 0xe4, 0x6f, 0xab, 0xc6, 0x6f, 0x9c, 0xe2, 0xb9, 0x60, 0x78, 0x76, 0x49, 0x89, 0x51, 0x30, 0x23, 0x1f, 0xa2, 0x62, 0x44, 0x16, 0xa4, 0x23, 0x7c, 0x85, 0x18, 0x74, 0xb5, 0xf9, 0xe8, 0x1b, 0xc2, 0x3d, 0x0d, 0xd2, 0x51, 0x1b, 0x74, 0x8e, 0x9c, 0x67, 0x97, 0x9c, 0x7a, 0x3e, 0x60, 0xe6, 0x40, 0x1e, 0x44, 0xe3, 0xc4, 0x0d, 0x06, 0xae, 0x1f, 0x06, 0x4c, 0x17, 0xb7, 0x7d, 0xd1, 0xe2, 0x9c, 0xf5, 0x29, 0xab, 0x2a, 0xec, 0x4f, 0xb8, 0x96, 0x0b, 0xc5, 0xab, 0xc2, 0x17, 0xaa, 0xca, 0x59, 0x9d, 0xf2, 0xf1, 0x62, 0xf6, 0xe0, 0x6a, 0xcc, 0x22, 0xe6, 0x66, 0xaa, 0x89, 0xcb, 0xf6, 0x06, 0x01, 0x74, 0xde, 0xa9, 0x29, 0xab, 0x18, 0x06, 0x4a, 0x30, 0xdb, 0xc2, 0x84, 0x00, 0x0a, 0x0a, 0x30, 0x71, 0x43, 0x6a, 0x98, 0x9c, 0x06, 0xfe, 0x4f, 0x00, 0xad, 0x39, 0x7a, 0xb5, 0x9e, 0x05, 0x8a, 0x40, 0xdb, 0xc2, 0xbb, 0x04, 0xd0, 0xb9, 0x22, 0xb0, 0xc4, 0x28, 0xa7, 0x81, 0x9b, 0x04, 0xd0, 0xf5, 0x29, 0xf0, 0xb4, 0xca, 0x28, 0x27, 0x82, 0x5b, 0x04, 0x50, 0x54, 0x04, 0xda, 0x16, 0xba, 0x03, 0xeb, 0x1a, 0xa8, 0xae, 0xf1, 0x3d, 0x02, 0xe8, 0x82, 0xa3, 0x09, 0xd4, 0x5c, 0x50, 0x81, 0xda, 0x16, 0xbe, 0x4d, 0x00, 0xbd, 0x5c, 0x82, 0xda, 0x16, 0xba, 0x07, 0xd7, 0xa7, 0xe1, 0x15, 0xed, 0x3e, 0x01, 0x74, 0xcd, 0xd1, 0x1c, 0x7a, 0x3e, 0xa8, 0x82, 0x6d, 0x0b, 0x53, 0x02, 0x68, 0xbd, 0x0c, 0x16, 0x33, 0xcc, 0x34, 0x09, 0x3e, 0x25, 0xdc, 0x21, 0x80, 0x9a, 0x53, 0xe9, 0xc5, 0x90, 0x90, 0xaf, 0x5f, 0x8e, 0x09, 0x77, 0x09, 0xa0, 0xc6, 0xb4, 0x7e, 0x39, 0x23, 0xdc, 0x84, 0x7a, 0xa5, 0x98, 0x12, 0xee, 0x11, 0x40, 0x17, 0x9d, 0x15, 0x65, 0xe4, 0x13, 0x42, 0x41, 0x4d, 0x31, 0x23, 0xdc, 0x27, 0x80, 0x2e, 0xe5, 0xd4, 0x14, 0xe3, 0x41, 0x3e, 0x3b, 0x31, 0x20, 0x1c, 0x10, 0x40, 0x57, 0xa6, 0xd9, 0x89, 0xe1, 0x20, 0x86, 0xdb, 0x1a, 0x56, 0x3a, 0x15, 0x5a, 0x04, 0x5c, 0xb0, 0x71, 0xb6, 0x14, 0x75, 0xf1, 0x54, 0xf0, 0xe0, 0x66, 0x39, 0x26, 0xef, 0x20, 0x8b, 0x80, 0x8b, 0x74, 0x10, 0x2a, 0x46, 0xe3, 0x5d, 0x64, 0xc3, 0xad, 0x92, 0x5c, 0xdd, 0xc8, 0x63, 0x7d, 0x86, 0x1f, 0x66, 0xa2, 0x75, 0xcc, 0x45, 0xd3, 0xd9, 0x28, 0x0a, 0xf7, 0x32, 0x73, 0x67, 0x23, 0xb1, 0x5e, 0xd7, 0x0f, 0xe3, 0x01, 0xb6, 0x25, 0xde, 0x98, 0xbe, 0x8f, 0x9f, 0xc2, 0x78, 0x80, 0xe2, 0x5c, 0x00, 0xdf, 0xfd, 0x3c, 0xd1, 0xea, 0x3d, 0xba, 0xa8, 0x7a, 0x1d, 0x93, 0x1a, 0xd3, 0xe4, 0x7e, 0x75, 0x3f, 0x4f, 0x94, 0x7e, 0x23, 0xb8, 0x3e, 0x72, 0x23, 0x55, 0x8f, 0xdc, 0x05, 0x3f, 0xf0, 0x78, 0x47, 0xe7, 0x8a, 0xf7, 0xdc, 0x8d, 0x44, 0xd1, 0xe2, 0xef, 0xd3, 0x20, 0x89, 0x27, 0xce, 0xda, 0xa8, 0x68, 0x45, 0x1f, 0xe1, 0x56, 0x16, 0x4e, 0xdc, 0xe1, 0xf2, 0x7d, 0x65, 0x9f, 0x60, 0xf8, 0x47, 0x1e, 0xf2, 0xc9, 0x79, 0x43, 0xf2, 0x13, 0x49, 0x94, 0x9a, 0xd9, 0x44, 0x54, 0x34, 0xaa, 0x38, 0x76, 0x6e, 0xc0, 0x5a, 0x71, 0xe3, 0xac, 0x42, 0xb3, 0xd7, 0x13, 0xdf, 0x45, 0x8e, 0xd9, 0xeb, 0xed, 0x74, 0xe0, 0xe6, 0x97, 0x4a, 0x40, 0x75, 0x08, 0xde, 0xb1, 0x09, 0x07, 0x2e, 0x39, 0xd9, 0x4f, 0xb4, 0x09, 0xe7, 0xc5, 0xf8, 0x67, 0x72, 0x9b, 0x78, 0x68, 0x9b, 0x8f, 0x8d, 0x9d, 0x09, 0xdc, 0x9e, 0x91, 0x53, 0x9e, 0x06, 0x08, 0x9a, 0x67, 0x79, 0x9a, 0xe5, 0x66, 0xf3, 0xbf, 0x4a, 0x97, 0x8c, 0x39, 0x01, 0x72, 0xa1, 0x77, 0xf7, 0x21, 0xcc, 0x6d, 0xd5, 0x05, 0x08, 0x8e, 0x5f, 0xbc, 0xa8, 0x5f, 0xca, 0x7e, 0x74, 0x8e, 0x9c, 0xba, 0x21, 0x7e, 0xbc, 0xa9, 0x9b, 0x9d, 0x1a, 0x5c, 0xce, 0xe9, 0xbe, 0xfb, 0xb7, 0x01, 0x51, 0x95, 0x39, 0x4b, 0xae, 0xff, 0xd6, 0xf3, 0x07, 0x3c, 0xe1, 0x6f, 0x4c, 0x8e, 0x13, 0xa0, 0x63, 0xb8, 0x10, 0xb9, 0x13, 0x3f, 0x74, 0x07, 0xb2, 0xd0, 0xfb, 0xe7, 0x79, 0xc7, 0x8e, 0x5a, 0xbc, 0xfb, 0x50, 0x7c, 0xe1, 0x3a, 0xec, 0x7d, 0xea, 0xc5, 0x6c, 0x20, 0xce, 0x24, 0xf1, 0x5b, 0x7f, 0xe3, 0x9a, 0xe2, 0x4e, 0x13, 0x56, 0x7e, 0xa7, 0xdd, 0xdd, 0x87, 0xf0, 0xc4, 0x0f, 0x7b, 0xae, 0xaf, 0x74, 0x39, 0x51, 0xba, 0x9c, 0x28, 0x5d, 0x4e, 0xb8, 0x2e, 0xed, 0x37, 0x8f, 0x87, 0x5e, 0xf2, 0x36, 0xed, 0x35, 0xfa, 0xe1, 0xe8, 0x50, 0x7e, 0x6f, 0xf7, 0x99, 0x7f, 0x30, 0x8e, 0x58, 0xff, 0xeb, 0xff, 0x3e, 0xf8, 0x37, 0x00, 0x00, 0xff, 0xff, 0x21, 0x77, 0xe0, 0x3d, 0x64, 0x10, 0x00, 0x00, }
<reponame>aaeabdo/coding-challenge-1 require 'spec_helper' require_relative '../../../logic/package/creator' RSpec.describe Logic::Package::Creator do subject(:call) { described_class.call(model, creation_params) } let(:model) { class_double('Package') } let(:creation_params) do { name: "ABC.RAP", version: "0.9.0", publication_date: '2016-10-20 10:52:16', title: "Array Based CpG Region Analysis Pipeline", description: "It aims to identify candidate genes that are “differentially methylated”", authors: [ { name: "<NAME>", email: nil, roles: ["cre", "aut"] } ], maintainers:[ { name: "<NAME>", email: "<EMAIL>", roles: ["cre"] } ] } end context '.call' do let(:package) { instance_double('Package') } let(:initialization_params) do creation_params.slice(:name, :version) end it 'runs successfully' do expect(model) .to receive(:find_or_initialize_by) .with(initialization_params) .and_return(package) expect(package) .to receive(:assign_attributes) .with(creation_params) expect(package) .to receive(:save!) call end end end
#!/bin/bash echo "=========================================================" date cd /home/pi/prog/garden_pi/utils sudo PYTHONPATH=/home/pi/prog/garden_pi ./water.py
#!/bin/bash ROOT_DIR=$(dirname $(dirname $(realpath "$0"))) find ${ROOT_DIR} -name Manifest -exec grep ^DIST "{}" \; \ | awk '{print$7" *"$2}' \ | (cd /usr/portage/distfiles/ && sha512sum -c)
import { Meteor } from 'meteor/meteor'; import { settings } from '../../settings'; Meteor.startup(function() { settings.add('AutoTranslate_Enabled', false, { type: 'boolean', group: 'Message', section: 'AutoTranslate', public: true }); settings.add('AutoTranslate_GoogleAPIKey', '', { type: 'string', group: 'Message', section: 'AutoTranslate', enableQuery: { _id: 'AutoTranslate_Enabled', value: true } }); });
<reponame>Xi-Plus/OJ-Code // By KRT girl xiplus #include <bits/stdc++.h> #define endl '\n' using namespace std; struct Node{ int visit,low; vector<int> son; }node[110]; int ans; void dfs(int i,int p,int d){ node[i].visit=node[i].low=d; bool ap=false; int child=0; for(int s:node[i].son){ if(s==p)continue; if(node[s].visit){ node[i].low=min(node[i].low,node[s].visit); } else { child++; dfs(s,i,d+1); node[i].low=min(node[i].low,node[s].low); if(node[s].low>=node[i].visit)ap=true; } } if((i==p&&child>1)||(i!=p&&ap)){ ans++; } } int main(){ // ios::sync_with_stdio(false); // cin.tie(0); int n,p,t; string s; while(cin>>n,n){ for(int q=1;q<=n;q++){ node[q].son.clear(); node[q].visit=0; } ans=0; while(cin>>p,p){ cin.get(); getline(cin,s); stringstream ss(s); while(ss>>t){ node[p].son.push_back(t); node[t].son.push_back(p); } } dfs(1,1,1); cout<<ans<<endl; } }
#!/bin/sh # Copyright 2020 Google LLC # # Licensed under the Apache License, Version 2.0 (the "License"); # you may not use this file except in compliance with the License. # You may obtain a copy of the License at # # http://www.apache.org/licenses/LICENSE-2.0 # # Unless required by applicable law or agreed to in writing, software # distributed under the License is distributed on an "AS IS" BASIS, # WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied. # See the License for the specific language governing permissions and # limitations under the License. network="default" subnet="default" #** # @brief Displays usage details. # usage() { echo -e "$*\n usage: $(basename "$0")" \ "-o <org> -e <env> -c <component> -n <namespace>\n" \ "example: $(basename "$0") -p my-proj -r us-west1 -i 10.75.0.2 \n" \ "Parameters:\n" \ "-p --prj : GCP Project Id (mandatory parameter)\n" \ "-r --reg : GCP Region Name (mandatory parameter)\n" \ "-i --ip : Apigee Instance IP (mandatory parameter)\n" \ "-n --network : Network name (optional parameter; default is default)\n" \ "-s --subnet : Subnet name (optional parameter; default is default)\n" exit 1 } ### Start of mainline code ### PARAMETERS=() while [[ $# -gt 0 ]] do param="$1" case $param in -p|--prj) project="$2" shift shift ;; -r|--reg) region="$2" shift shift ;; -i|--ip) apigeeip="$2" shift shift ;; -n|--network) network="$2" shift shift ;; -s|--subnet) subnet="$2" shift shift ;; *) PARAMETERS+=("$1") shift ;; esac done set -- "${PARAMETERS[@]}" ./check-prereqs.sh $project $region $apigeeip RESULT=$? if [ $RESULT -ne 0 ]; then usage exit 1 fi ./setup-mig.sh $project $region $apigeeip $network $subnet RESULT=$? if [ $RESULT -ne 0 ]; then exit 1 fi while true; do read -p "Do you to proceed with the creation and configuration of GCLB?" yn case $yn in [Yy]* ) break;; [Nn]* ) exit 0;; * ) echo "Please enter yes or no.";; esac done ./setup-loadbalancer.sh $project $region $network $subnet RESULT=$? if [ $RESULT -ne 0 ]; then exit 1 fi exit 0
CREATE TABLE version_owner_actions ( id SERIAL PRIMARY KEY, version_id INTEGER REFERENCES versions(id) ON DELETE CASCADE, owner_id INTEGER REFERENCES users(id), owner_token_id INTEGER REFERENCES api_tokens(id), action INTEGER NOT NULL, time TIMESTAMP NOT NULL DEFAULT now() );
import { Component, OnInit, ChangeDetectionStrategy, ChangeDetectorRef } from '@angular/core'; import { interval, pipe } from 'rxjs'; import { map, startWith } from 'rxjs/operators'; import { UntilDestroy, untilDestroyed } from '@ngneat/until-destroy'; import { JanusService } from '@shared/services/janus.service'; import { StateService } from '@shared/services/state.service'; import { LayoutType } from '@enums'; import { CustomizationModel } from '../../models/customization.model'; import { CustomizationService } from '../../services/customization.service'; @UntilDestroy() @Component({ selector: 'stusan-header', templateUrl: './header.component.html', styleUrls: ['./header.component.scss'], changeDetection: ChangeDetectionStrategy.OnPush, }) export class HeaderComponent implements OnInit { public roomName?: string; public isSirius = false; public customConfig: CustomizationModel; constructor( public janusService: JanusService, public stateService: StateService, public customizationService: CustomizationService, private cdr: ChangeDetectorRef, ) { this.isSirius = customizationService.isSirius(); this.customConfig = customizationService.config; } public startTimer$ = interval(1000).pipe( map((seconds: number) => { if (this.stateService.startTime) { return new Date(Date.now() - this.stateService.startTime) .toUTCString() .substring(17, 25); } else { return new Date(); } // In order to get rid of time offset }), startWith('00:00:00'), pipe(untilDestroyed(this)), ); public LayoutTypeEnum = LayoutType; ngOnInit(): void { this.stateService.roomName$.subscribe((name) => { if (name) { this.roomName = name; } this.cdr.markForCheck(); }); } public setLayoutType(layoutType: LayoutType): void { this.stateService.layoutType$.next(layoutType); } }
#!/usr/bin/env bash set -e # exit on errors BASE_DIR="$( cd "$(dirname "$0")" ; pwd -P )" # Folders in which whitesource has to be run declare -a FOLDERS=("client" "core" "website/landingpage/dev" "website/fiddle" "plugins" ) # Project tokens for corresponding folders declare -a PROJECT_TOKENS=("a63fd5aaaa2343199327aac6d3e2b5346e930927d66441bf92111f111ecfc8ad" "dd9b33e88b684129b325ef1c3510f52712f26167f06c432ab84fd4452cd353d9" "0bc00859083a4cf98e1246c91cabc56a4623fb69f1f24cdd8ca10f9f0d3dfbb4" "8b3ce944979d49c8b1d2917533199aeb4e73a993843f4820a509a33932e18a29" "56eb623145264763a82330025e4f11efd538074a3cd04791ab6ab87293174aca" ) if [ -f "$CACHED_WHITESOURCE_UA" ]; then echo "$CACHED_WHITESOURCE_UA exists. Take from Cache." else # download jar file to /home/.cache/whitesource cd $BASE_DIR/../../../../.cache/whitesource CACHED_WHITESOURCE_UA=$BASE_DIR/../../../../.cache/whitesource/wss-unified-agent.jar echo "$CACHED_WHITESOURCE_UA does not exist." echo "Start download.." curl -LJO https://github.com/whitesource/unified-agent-distribution/releases/latest/download/wss-unified-agent.jar echo "Downloaded" fi for ((i=0;i<${#FOLDERS[@]};++i)); do cd $BASE_DIR/../${FOLDERS[i]} echo "Create config File:" echo $'apiKey='${WHITESOURCE_APIKEY}$'\nuserKey='$WHITESOURCE_USER_TOKEN$'\nproductName='$WHITESOURCE_PRODUCT_TOKEN$'\nprojectToken='${PROJECT_TOKENS[i]}$'\ndevDep=false\nforceUpdate=true\ncheckPolicies=true\nwss.url=https://sap.whitesourcesoftware.com/agent' > wss-generated-file.config java -jar $BASE_DIR/../../../../.cache/whitesource/wss-unified-agent.jar -c wss-generated-file.config -d . -scanComment "$(date)" RV=$? echo "Exit code: $RV" done exit 0
<reponame>multiplex/multiplex.js<filename>src/lib/collections/stack.js import Collection from './collection'; import buffer from '../utils/buffer'; import extend from '../utils/extend'; import iterableSymbol from '../iteration/iterable-symbol'; import error, {ERROR_EMPTY_COLLECTION} from '../utils/error'; /** * Initializes a new instance of the Stack class that that is empty or contains elements copied from the specified collection. * @param {Iterable=} collection The collection to copy elements from. */ export default function Stack(collection) { var items = collection ? buffer(collection) : []; Collection.call(this, items); } extend(Stack, Collection, { /** * Removes all objects from the Stack. */ clear: function () { this[iterableSymbol].length = 0; }, /** * Determines whether an element is in the Stack. * @param {Object} item The object to locate in the Stack. * @returns {Boolean} */ contains: function (item) { return this[iterableSymbol].indexOf(item) !== -1; }, /** * Returns the object at the top of the Stack without removing it. * @returns {Object} */ peek: function () { var items = this[iterableSymbol]; if (items.length > 0) { return items[items.length - 1]; } error(ERROR_EMPTY_COLLECTION); }, /** * Removes and returns the object at the top of the Stack. * @returns {Object} */ pop: function () { if (this.count() > 0) { return this[iterableSymbol].pop(); } error(ERROR_EMPTY_COLLECTION); }, /** * Inserts an object at the top of the Stack. * @param {Object} item The object to push onto the Stack. */ push: function (item) { this[iterableSymbol].push(item); }, toString: function () { return '[Stack]'; } });
package pl.allegro.tech.boot.leader.only; import org.springframework.beans.factory.config.BeanPostProcessor; import org.springframework.lang.Nullable; import pl.allegro.tech.boot.leader.only.api.Leader; import static org.springframework.core.annotation.AnnotationUtils.findAnnotation; final class LeaderOnlyBeanPostProcessor implements BeanPostProcessor { private final LeadershipProxyFactory leadershipProxyFactory; public LeaderOnlyBeanPostProcessor(LeadershipProxyFactory leadershipProxyFactory) { this.leadershipProxyFactory = leadershipProxyFactory; } @Override public Object postProcessAfterInitialization(Object bean, @Nullable String beanName) { Leader annotation = findAnnotation(bean.getClass(), Leader.class); if (annotation == null) { return bean; } return leadershipProxyFactory.getProxy(bean, annotation.value()); } }
import numpy as np import tensorflow as tf from tensorflow.keras.preprocessing.sequence import pad_sequences from tensorflow.keras.layers import Dense, Embedding, LSTM from tensorflow.keras.models import Sequential from tensorflow.keras.optimizers import Adam vocab = ['This', 'is', 'a', 'sentence', 'that', 'the', 'model', 'will', 'predict', 'the', 'next', 'word', 'of'] max_len = 4 train_data = [ [vocab[0], vocab[1], vocab[2], vocab[3]], [vocab[1], vocab[2], vocab[3], vocab[4]], [vocab[2], vocab[3], vocab[4], vocab[5]], [vocab[3], vocab[4], vocab[5], vocab[6]], [vocab[4], vocab[5], vocab[6], vocab[7]], [vocab[5], vocab[6], vocab[7], vocab[8]], [vocab[6], vocab[7], vocab[8], vocab[9]], [vocab[7], vocab[8], vocab[9], vocab[10]], [vocab[8], vocab[9], vocab[10], vocab[11]] ] # Convert data to binary matrix training_data = [[vocab.index(word) for word in lst] for lst in train_data] # Pad the matrix to max_len training_data = pad_sequences(training_data, maxlen=max_len, dtype='float64') # Split the input into 4 features x_train = np.array([training_data[:, 0], training_data[:, 1], training_data[:, 2], training_data[:, 3]]).T # Create the target array y_train = np.roll(training_data, -1, axis=1) # Build the RNN model model = Sequential() model.add(Embedding(input_dim=len(vocab)+1, output_dim=32, input_length=4)) model.add(LSTM(250)) model.add(Dense(len(vocab), activation='softmax')) model.compile(loss='categorical_crossentropy', optimizer=Adam(learning_rate=0.1)) # Train the model model.fit(x_train, y_train, epochs=200, verbose=1)
var SongFlag = 0 var SongScale = 1 function NewSongQuery() { for (i = 1; i <= 16; i++) { let tr = tab.tHead.children[0], tr2 = document.getElementById('tbm').children[0], th = document.createElement('th'), td = document.createElement('td'); th.innerHTML = tab.rows[0].cells.length + 1; th.setAttribute("onclick","CallJump(this.innerHTML)") td.classList.add("editMe") td.innerHTML = '' tr.appendChild(th); tr2.appendChild(td); } // open loop for each row and append cell } function create_hell_song() { const data = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/"; const notes = "BRhJCSKEUkiM"; const totalValues = Math.pow(data.length, 2); let TempOutput=Math.floor(Math.random() * 7) + 1 let output = TempOutput.toString() for (let i = 0; i < 100; i++) { output += notes[Math.floor(Math.random() * notes.length)]; const c1 = data[Math.floor(i / data.length)]; const c2 = data[i % data.length]; output += c1 + c2; } console.log(output); document.getElementById("HellSongString").value = output const input = document.querySelector('#HellSongString'); input.select(); if (document.execCommand('copy')) { document.execCommand('copy'); console.log('複製成功'); $("[data-toggle='HSSpopover']").popover("toggle"); setTimeout(function () { $("[data-toggle='HSSpopover']").popover('hide'); }, 2000); } } function copySong() { const input = document.querySelector('#SongString'); input.select(); if (document.execCommand('copy')) { document.execCommand('copy'); console.log('複製成功'); $("[data-toggle='SSpopover']").popover("toggle"); setTimeout(function () { $("[data-toggle='SSpopover']").popover('hide'); }, 2000); } } function FlagMove(direction, moveNum) { if (direction == "-" && SongFlag != 0) { tab.rows[0].cells[SongFlag].style.background = "#343a40"; SongFlag--; document.getElementById("FlagJump").value-- tab.rows[0].cells[SongFlag].style.background = "red"; } else if (direction == "+" && SongFlag != tab.rows[0].cells.length - 1) { tab.rows[0].cells[SongFlag].style.background = "#343a40"; SongFlag++; document.getElementById("FlagJump").value++ tab.rows[0].cells[SongFlag].style.background = "red"; } } function CallJump(num){ document.getElementById("FlagJump").value=num FJump("FlagJump") } function FJump(e) { let tempF = document.getElementById(e).value if (tempF >= tab.rows[0].cells.length - 1) { let tableng = tab.rows[0].cells.length tab.rows[0].cells[SongFlag].style.background = "#343a40"; SongFlag = tableng - 1; tab.rows[0].cells[SongFlag].style.background = "red"; document.getElementById(e).value = tableng } else { tab.rows[0].cells[SongFlag].style.background = "#343a40"; SongFlag = tempF - 1; tab.rows[0].cells[SongFlag].style.background = "red"; } } function ScaleChange(SN) { if (SN != SongScale) { var NT = document.getElementById('NameTable') document.getElementById("ScaleMenu").innerHTML = document.getElementById("SC" + SN).innerHTML document.getElementById("SC" + SongScale).classList.remove("active") document.getElementById("SC" + SN).classList.add("active") SongScale = SN $("#NameTable").fadeOut("normal", function () { NT.tHead.children[0].innerHTML = "<th colspan=2>" + document.getElementById("SC" + SN).innerHTML + "</th>" var TBFlag = 2 for (i = 11; i >= 0; i--) { NT.rows[TBFlag].cells[1].innerText = notename[SN][i] TBFlag++ } }); $("#NameTable").fadeIn("normal"); } } function SongInput(input){ tab.rows[1].cells[SongFlag].innerHTML=input OutputString() } function OutputString(){ var String = SongScale const data = "ABCDEFGHIJKLMNOPQRSTUVWXYZabcdefghijklmnopqrstuvwxyz0123456789+/"; for(i=0;i<tab.rows[0].cells.length;i++){ if(tab.rows[1].cells[i].innerHTML!=""){ const c1 = data[Math.floor(i / data.length)]; const c2 = data[i % data.length]; String += tab.rows[1].cells[i].innerHTML+c1+c2 } } document.getElementById('SongString').value=String console.log(String) } function SongClear(){ for(i=0;i<tab.rows[0].cells.length;i++){ tab.rows[1].cells[i].innerHTML = "" } document.getElementById('SongString').value="" }
<filename>Mosaic Decoration II/main.cpp #include <iostream> using namespace std; unsigned long long w,h,a,b,m,c,res=0; int main() { cin >> w >> h >> a >> b >> m >> c; unsigned long long int maxRight = w / a + (w%a==0 ? 0 : 1); unsigned long long int maxBottom = h / b + (h%b==0 ? 0 : 1); unsigned long long int cutsRight = (w % a != 0 ? h : 0); // cout << cutsRight << endl; unsigned long long int cutsBottom = (h % b != 0 ? w : 0); // cout << cutsBottom << endl; unsigned long long int totalCuts = cutsBottom + cutsRight ; unsigned long long int tilesSum = maxBottom*maxRight; unsigned long long int res = (tilesSum/10+(tilesSum%10 == 0 ? 0 : 1))*m; res += totalCuts*c; cout << res << endl; return 0; }
const request = require('supertest'); const server = require('../server'); const db = require('../../data/db-config'); describe('articles router', () => { it('does not return data unless there is a valid JSON web token in the header', async () => { const res = await request(server).get('/api/articles'); expect(res.body).toMatchObject({ message: 'token required' }); }); }); // the test passes, but cannot be used with NODE_ENV. DB_ENV ruins the rest of the server, so it cannot be used.
<gh_stars>0 /// <reference path="../common/models.ts" /> /// <reference path="../common/messaging.ts" /> /// <reference path="config.ts" /> /// <reference path="utils.ts" /> /// <reference path="interfaces.ts"/> /// <reference path="quoter.ts"/> /// <reference path="safety.ts"/> /// <reference path="statistics.ts"/> /// <reference path="active-state.ts"/> /// <reference path="fair-value.ts"/> /// <reference path="market-filtration.ts"/> /// <reference path="quoting-parameters.ts"/> /// <reference path="position-management.ts"/> /// <reference path="./quoting-styles/style-registry.ts"/> import Config = require("./config"); import Models = require("../common/models"); import Messaging = require("../common/messaging"); import Utils = require("./utils"); import Interfaces = require("./interfaces"); import Quoter = require("./quoter"); import Safety = require("./safety"); import util = require("util"); import _ = require("lodash"); import Statistics = require("./statistics"); import Active = require("./active-state"); import FairValue = require("./fair-value"); import MarketFiltration = require("./market-filtration"); import QuotingParameters = require("./quoting-parameters"); import PositionManagement = require("./position-management"); import moment = require('moment'); import QuotingStyleRegistry = require("./quoting-styles/style-registry"); import {QuoteInput} from "./quoting-styles/helpers"; import log from "./logging"; export class QuotingEngine { private _log = log("quotingengine"); public QuoteChanged = new Utils.Evt<Models.TwoSidedQuote>(); private _latest: Models.TwoSidedQuote = null; public get latestQuote() { return this._latest; } public set latestQuote(val: Models.TwoSidedQuote) { if (!quotesChanged(this._latest, val, this._details.minTickIncrement)) return; this._latest = val; this.QuoteChanged.trigger(); this._quotePublisher.publish(this._latest); } constructor( private _registry: QuotingStyleRegistry.QuotingStyleRegistry, private _timeProvider: Utils.ITimeProvider, private _filteredMarkets: MarketFiltration.MarketFiltration, private _fvEngine: FairValue.FairValueEngine, private _qlParamRepo: QuotingParameters.QuotingParametersRepository, private _quotePublisher: Messaging.IPublish<Models.TwoSidedQuote>, private _orderBroker: Interfaces.IOrderBroker, private _positionBroker: Interfaces.IPositionBroker, private _details: Interfaces.IBroker, private _ewma: Interfaces.IEwmaCalculator, private _targetPosition: PositionManagement.TargetBasePositionManager, private _safeties: Safety.SafetyCalculator) { var recalcWithoutInputTime = () => this.recalcQuote(_timeProvider.utcNow()); _filteredMarkets.FilteredMarketChanged.on(m => this.recalcQuote(Utils.timeOrDefault(m, _timeProvider))); _qlParamRepo.NewParameters.on(recalcWithoutInputTime); _orderBroker.Trade.on(recalcWithoutInputTime); _ewma.Updated.on(recalcWithoutInputTime); _quotePublisher.registerSnapshot(() => this.latestQuote === null ? [] : [this.latestQuote]); _targetPosition.NewTargetPosition.on(recalcWithoutInputTime); _safeties.NewValue.on(recalcWithoutInputTime); _timeProvider.setInterval(recalcWithoutInputTime, moment.duration(1, "seconds")); } private computeQuote(filteredMkt: Models.Market, fv: Models.FairValue) { const params = this._qlParamRepo.latest; const minTick = this._details.minTickIncrement; const input = new QuoteInput(filteredMkt, fv, params, minTick); const unrounded = this._registry.Get(params.mode).GenerateQuote(input); if (unrounded === null) return null; if (params.ewmaProtection && this._ewma.latest !== null) { if (this._ewma.latest > unrounded.askPx) { unrounded.askPx = Math.max(this._ewma.latest, unrounded.askPx); } if (this._ewma.latest < unrounded.bidPx) { unrounded.bidPx = Math.min(this._ewma.latest, unrounded.bidPx); } } const tbp = this._targetPosition.latestTargetPosition; if (tbp === null) { this._log.warn("cannot compute a quote since no position report exists!"); return null; } const targetBasePosition = tbp.data; const latestPosition = this._positionBroker.latestReport; const totalBasePosition = latestPosition.baseAmount + latestPosition.baseHeldAmount; if (totalBasePosition < targetBasePosition - params.positionDivergence) { unrounded.askPx = null; unrounded.askSz = null; if (params.aggressivePositionRebalancing) unrounded.bidSz = Math.min(params.aprMultiplier*params.size, targetBasePosition - totalBasePosition); } if (totalBasePosition > targetBasePosition + params.positionDivergence) { unrounded.bidPx = null; unrounded.bidSz = null; if (params.aggressivePositionRebalancing) unrounded.askSz = Math.min(params.aprMultiplier*params.size, totalBasePosition - targetBasePosition); } const safety = this._safeties.latest; if (safety === null) { return null; } if (params.mode === Models.QuotingMode.PingPong) { if (unrounded.askSz && safety.buyPing && unrounded.askPx < safety.buyPing + params.width) unrounded.askPx = safety.buyPing + params.width; if (unrounded.bidSz && safety.sellPong && unrounded.bidPx > safety.sellPong - params.width) unrounded.bidPx = safety.sellPong - params.width; } if (safety.sell > params.tradesPerMinute) { unrounded.askPx = null; unrounded.askSz = null; } if (safety.buy > params.tradesPerMinute) { unrounded.bidPx = null; unrounded.bidSz = null; } if (unrounded.bidPx !== null) { unrounded.bidPx = Utils.roundSide(unrounded.bidPx, minTick, Models.Side.Bid); unrounded.bidPx = Math.max(0, unrounded.bidPx); } if (unrounded.askPx !== null) { unrounded.askPx = Utils.roundSide(unrounded.askPx, minTick, Models.Side.Ask); unrounded.askPx = Math.max(unrounded.bidPx + minTick, unrounded.askPx); } if (unrounded.askSz !== null) { unrounded.askSz = Utils.roundDown(unrounded.askSz, minTick); unrounded.askSz = Math.max(minTick, unrounded.askSz); } if (unrounded.bidSz !== null) { unrounded.bidSz = Utils.roundDown(unrounded.bidSz, minTick); unrounded.bidSz = Math.max(minTick, unrounded.bidSz); } return unrounded; } private recalcQuote = (t: Date) => { const fv = this._fvEngine.latestFairValue; if (fv == null) { this.latestQuote = null; return; } const filteredMkt = this._filteredMarkets.latestFilteredMarket; if (filteredMkt == null) { this.latestQuote = null; return; } const genQt = this.computeQuote(filteredMkt, fv); if (genQt === null) { this.latestQuote = null; return; } this.latestQuote = new Models.TwoSidedQuote( this.quotesAreSame(new Models.Quote(genQt.bidPx, genQt.bidSz), this.latestQuote, Models.Side.Bid), this.quotesAreSame(new Models.Quote(genQt.askPx, genQt.askSz), this.latestQuote, Models.Side.Ask), t ); }; private quotesAreSame( newQ: Models.Quote, prevTwoSided: Models.TwoSidedQuote, side: Models.Side): Models.Quote { if (newQ.price === null && newQ.size === null) return null; if (prevTwoSided == null) return newQ; const previousQ = Models.Side.Bid === side ? prevTwoSided.bid : prevTwoSided.ask; if (previousQ == null && newQ != null) return newQ; if (Math.abs(newQ.size - previousQ.size) > 5e-3) return newQ; if (Math.abs(newQ.price - previousQ.price) < this._details.minTickIncrement) { return previousQ; } let quoteWasWidened = true; if (Models.Side.Bid === side && previousQ.price < newQ.price) quoteWasWidened = false; if (Models.Side.Ask === side && previousQ.price > newQ.price) quoteWasWidened = false; // prevent flickering if (!quoteWasWidened && Math.abs(Utils.fastDiff(new Date(), prevTwoSided.time)) < 300) { return previousQ; } return newQ; } } const quoteChanged = (o: Models.Quote, n: Models.Quote, tick: number) : boolean => { if ((!o && n) || (o && !n)) return true; if (!o && !n) return false; const oPx = (o && o.price) || 0; const nPx = (n && n.price) || 0; if (Math.abs(oPx - nPx) > tick) return true; const oSz = (o && o.size) || 0; const nSz = (n && n.size) || 0; return Math.abs(oSz - nSz) > .001; } const quotesChanged = (o: Models.TwoSidedQuote, n: Models.TwoSidedQuote, tick: number) : boolean => { if ((!o && n) || (o && !n)) return true; if (!o && !n) return false; if (quoteChanged(o.bid, n.bid, tick)) return true; if (quoteChanged(o.ask, n.ask, tick)) return true; return false; }
<filename>src/modules/Balance/Balance.tsx import * as React from 'react'; import { compose } from 'redux'; import { Route, Switch } from 'react-router-dom'; import { userIsLogged, userAcceptedTOS, userConfirmedSecurityNotice } from 'modules/shared/checkAuth'; import { Module } from 'shared/types/app'; import { layouts } from './view/layouts'; import { ConfirmWithdrawalLayout } from './view/containers'; import { routes } from './constants'; const userIsCleared = compose(userIsLogged, userAcceptedTOS, userConfirmedSecurityNotice); class BalanceModule extends Module { public getRoutes() { return ( <Route key={routes.balance.getElementKey()} path={routes.balance.getPath()} > <Switch> <Route key={routes.balance['confirm-withdrawal'].getElementKey()} path={routes.balance['confirm-withdrawal'].getPath()} component={userIsCleared(ConfirmWithdrawalLayout as any)} /> <Route key={routes.balance.getElementKey()} path={routes.balance.getPath()} component={userIsCleared(layouts[routes.balance.getPath()] as any)} /> </Switch> </Route> ); } } export default BalanceModule;
#!/bin/bash # parse command-line options while [ "$1" != "" ]; do case $1 in -a | --all ) shift ALL=1 ;; * ) exit 1 esac shift done FILES='' if [[ ${ALL} ]]; then # compile all echo "Compiling all *.less files" FILES=`find . -name "*.less"` else # compile deltas echo "Compiling modified *.less files" FILES=`git status | grep "\.less" | awk '{ print $3 }'` fi for INFILE in $FILES do echo "Compiling ${INFILE}" FILENAME=${INFILE%.*} OUTFILE=${FILENAME}.css lessc ${INFILE} ${OUTFILE} done
#! /bin/sh ## DO NOT EDIT - This file generated from ./build-aux/ltmain.in ## by inline-source v2014-01-03.01 # libtool (GNU libtool) 2.4.6 # Provide generalized library-building support services. # Written by Gordon Matzigkeit <gord@gnu.ai.mit.edu>, 1996 # Copyright (C) 1996-2015 Free Software Foundation, Inc. # Copyright (C) 2018 The QpayCoin developers # This is free software; see the source for copying conditions. There is NO # warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # GNU Libtool is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 2 of the License, or # (at your option) any later version. # # As a special exception to the GNU General Public License, # if you distribute this file as part of a program or library that # is built using GNU Libtool, you may include this file under the # same distribution terms that you use for the rest of that program. # # GNU Libtool is distributed in the hope that it will be useful, but # WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. PROGRAM=libtool PACKAGE=libtool VERSION="2.4.6 Debian-2.4.6-2" package_revision=2.4.6 ## ------ ## ## Usage. ## ## ------ ## # Run './libtool --help' for help with using this script from the # command line. ## ------------------------------- ## ## User overridable command paths. ## ## ------------------------------- ## # After configure completes, it has a better idea of some of the # shell tools we need than the defaults used by the functions shared # with bootstrap, so set those here where they can still be over- # ridden by the user, but otherwise take precedence. : ${AUTOCONF="autoconf"} : ${AUTOMAKE="automake"} ## -------------------------- ## ## Source external libraries. ## ## -------------------------- ## # Much of our low-level functionality needs to be sourced from external # libraries, which are installed to $pkgauxdir. # Set a version string for this script. scriptversion=2015-01-20.17; # UTC # General shell script boiler plate, and helper functions. # Written by Gary V. Vaughan, 2004 # Copyright (C) 2004-2015 Free Software Foundation, Inc. # This is free software; see the source for copying conditions. There is NO # warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # This program is free software; you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation; either version 3 of the License, or # (at your option) any later version. # As a special exception to the GNU General Public License, if you distribute # this file as part of a program or library that is built using GNU Libtool, # you may include this file under the same distribution terms that you use # for the rest of that program. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNES FOR A PARTICULAR PURPOSE. See the GNU # General Public License for more details. # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # Please report bugs or propose patches to gary@gnu.org. ## ------ ## ## Usage. ## ## ------ ## # Evaluate this file near the top of your script to gain access to # the functions and variables defined here: # # . `echo "$0" | ${SED-sed} 's|[^/]*$||'`/build-aux/funclib.sh # # If you need to override any of the default environment variable # settings, do that before evaluating this file. ## -------------------- ## ## Shell normalisation. ## ## -------------------- ## # Some shells need a little help to be as Bourne compatible as possible. # Before doing anything else, make sure all that help has been provided! DUALCASE=1; export DUALCASE # for MKS sh if test -n "${ZSH_VERSION+set}" && (emulate sh) >/dev/null 2>&1; then : emulate sh NULLCMD=: # Pre-4.2 versions of Zsh do word splitting on ${1+"$@"}, which # is contrary to our usage. Disable this feature. alias -g '${1+"$@"}'='"$@"' setopt NO_GLOB_SUBST else case `(set -o) 2>/dev/null` in *posix*) set -o posix ;; esac fi # NLS nuisances: We save the old values in case they are required later. _G_user_locale= _G_safe_locale= for _G_var in LANG LANGUAGE LC_ALL LC_CTYPE LC_COLLATE LC_MESSAGES do eval "if test set = \"\${$_G_var+set}\"; then save_$_G_var=\$$_G_var $_G_var=C export $_G_var _G_user_locale=\"$_G_var=\\\$save_\$_G_var; \$_G_user_locale\" _G_safe_locale=\"$_G_var=C; \$_G_safe_locale\" fi" done # CDPATH. (unset CDPATH) >/dev/null 2>&1 && unset CDPATH # Make sure IFS has a sensible default sp=' ' nl=' ' IFS="$sp $nl" # There are apparently some retarded systems that use ';' as a PATH separator! if test "${PATH_SEPARATOR+set}" != set; then PATH_SEPARATOR=: (PATH='/bin;/bin'; FPATH=$PATH; sh -c :) >/dev/null 2>&1 && { (PATH='/bin:/bin'; FPATH=$PATH; sh -c :) >/dev/null 2>&1 || PATH_SEPARATOR=';' } fi ## ------------------------- ## ## Locate command utilities. ## ## ------------------------- ## # func_executable_p FILE # ---------------------- # Check that FILE is an executable regular file. func_executable_p () { test -f "$1" && test -x "$1" } # func_path_progs PROGS_LIST CHECK_FUNC [PATH] # -------------------------------------------- # Search for either a program that responds to --version with output # containing "GNU", or else returned by CHECK_FUNC otherwise, by # trying all the directories in PATH with each of the elements of # PROGS_LIST. # # CHECK_FUNC should accept the path to a candidate program, and # set $func_check_prog_result if it truncates its output less than # $_G_path_prog_max characters. func_path_progs () { _G_progs_list=$1 _G_check_func=$2 _G_PATH=${3-"$PATH"} _G_path_prog_max=0 _G_path_prog_found=false _G_save_IFS=$IFS; IFS=${PATH_SEPARATOR-:} for _G_dir in $_G_PATH; do IFS=$_G_save_IFS test -z "$_G_dir" && _G_dir=. for _G_prog_name in $_G_progs_list; do for _exeext in '' .EXE; do _G_path_prog=$_G_dir/$_G_prog_name$_exeext func_executable_p "$_G_path_prog" || continue case `"$_G_path_prog" --version 2>&1` in *GNU*) func_path_progs_result=$_G_path_prog _G_path_prog_found=: ;; *) $_G_check_func $_G_path_prog func_path_progs_result=$func_check_prog_result ;; esac $_G_path_prog_found && break 3 done done done IFS=$_G_save_IFS test -z "$func_path_progs_result" && { echo "no acceptable sed could be found in \$PATH" >&2 exit 1 } } # We want to be able to use the functions in this file before configure # has figured out where the best binaries are kept, which means we have # to search for them ourselves - except when the results are already set # where we skip the searches. # Unless the user overrides by setting SED, search the path for either GNU # sed, or the sed that truncates its output the least. test -z "$SED" && { _G_sed_script=s/aaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaaa/bbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbbb/ for _G_i in 1 2 3 4 5 6 7; do _G_sed_script=$_G_sed_script$nl$_G_sed_script done echo "$_G_sed_script" 2>/dev/null | sed 99q >conftest.sed _G_sed_script= func_check_prog_sed () { _G_path_prog=$1 _G_count=0 printf 0123456789 >conftest.in while : do cat conftest.in conftest.in >conftest.tmp mv conftest.tmp conftest.in cp conftest.in conftest.nl echo '' >> conftest.nl "$_G_path_prog" -f conftest.sed <conftest.nl >conftest.out 2>/dev/null || break diff conftest.out conftest.nl >/dev/null 2>&1 || break _G_count=`expr $_G_count + 1` if test "$_G_count" -gt "$_G_path_prog_max"; then # Best one so far, save it but keep looking for a better one func_check_prog_result=$_G_path_prog _G_path_prog_max=$_G_count fi # 10*(2^10) chars as input seems more than enough test 10 -lt "$_G_count" && break done rm -f conftest.in conftest.tmp conftest.nl conftest.out } func_path_progs "sed gsed" func_check_prog_sed $PATH:/usr/xpg4/bin rm -f conftest.sed SED=$func_path_progs_result } # Unless the user overrides by setting GREP, search the path for either GNU # grep, or the grep that truncates its output the least. test -z "$GREP" && { func_check_prog_grep () { _G_path_prog=$1 _G_count=0 _G_path_prog_max=0 printf 0123456789 >conftest.in while : do cat conftest.in conftest.in >conftest.tmp mv conftest.tmp conftest.in cp conftest.in conftest.nl echo 'GREP' >> conftest.nl "$_G_path_prog" -e 'GREP$' -e '-(cannot match)-' <conftest.nl >conftest.out 2>/dev/null || break diff conftest.out conftest.nl >/dev/null 2>&1 || break _G_count=`expr $_G_count + 1` if test "$_G_count" -gt "$_G_path_prog_max"; then # Best one so far, save it but keep looking for a better one func_check_prog_result=$_G_path_prog _G_path_prog_max=$_G_count fi # 10*(2^10) chars as input seems more than enough test 10 -lt "$_G_count" && break done rm -f conftest.in conftest.tmp conftest.nl conftest.out } func_path_progs "grep ggrep" func_check_prog_grep $PATH:/usr/xpg4/bin GREP=$func_path_progs_result } ## ------------------------------- ## ## User overridable command paths. ## ## ------------------------------- ## # All uppercase variable names are used for environment variables. These # variables can be overridden by the user before calling a script that # uses them if a suitable command of that name is not already available # in the command search PATH. : ${CP="cp -f"} : ${ECHO="printf %s\n"} : ${EGREP="$GREP -E"} : ${FGREP="$GREP -F"} : ${LN_S="ln -s"} : ${MAKE="make"} : ${MKDIR="mkdir"} : ${MV="mv -f"} : ${RM="rm -f"} : ${SHELL="${CONFIG_SHELL-/bin/sh}"} ## -------------------- ## ## Useful sed snippets. ## ## -------------------- ## sed_dirname='s|/[^/]*$||' sed_basename='s|^.*/||' # Sed substitution that helps us do robust quoting. It backslashifies # metacharacters that are still active within double-quoted strings. sed_quote_subst='s|\([`"$\\]\)|\\\1|g' # Same as above, but do not quote variable references. sed_double_quote_subst='s/\(["`\\]\)/\\\1/g' # Sed substitution that turns a string into a regex matching for the # string literally. sed_make_literal_regex='s|[].[^$\\*\/]|\\&|g' # Sed substitution that converts a w32 file name or path # that contains forward slashes, into one that contains # (escaped) backslashes. A very naive implementation. sed_naive_backslashify='s|\\\\*|\\|g;s|/|\\|g;s|\\|\\\\|g' # Re-'\' parameter expansions in output of sed_double_quote_subst that # were '\'-ed in input to the same. If an odd number of '\' preceded a # '$' in input to sed_double_quote_subst, that '$' was protected from # expansion. Since each input '\' is now two '\'s, look for any number # of runs of four '\'s followed by two '\'s and then a '$'. '\' that '$'. _G_bs='\\' _G_bs2='\\\\' _G_bs4='\\\\\\\\' _G_dollar='\$' sed_double_backslash="\ s/$_G_bs4/&\\ /g s/^$_G_bs2$_G_dollar/$_G_bs&/ s/\\([^$_G_bs]\\)$_G_bs2$_G_dollar/\\1$_G_bs2$_G_bs$_G_dollar/g s/\n//g" ## ----------------- ## ## Global variables. ## ## ----------------- ## # Except for the global variables explicitly listed below, the following # functions in the '^func_' namespace, and the '^require_' namespace # variables initialised in the 'Resource management' section, sourcing # this file will not pollute your global namespace with anything # else. There's no portable way to scope variables in Bourne shell # though, so actually running these functions will sometimes place # results into a variable named after the function, and often use # temporary variables in the '^_G_' namespace. If you are careful to # avoid using those namespaces casually in your sourcing script, things # should continue to work as you expect. And, of course, you can freely # overwrite any of the functions or variables defined here before # calling anything to customize them. EXIT_SUCCESS=0 EXIT_FAILURE=1 EXIT_MISMATCH=63 # $? = 63 is used to indicate version mismatch to missing. EXIT_SKIP=77 # $? = 77 is used to indicate a skipped test to automake. # Allow overriding, eg assuming that you follow the convention of # putting '$debug_cmd' at the start of all your functions, you can get # bash to show function call trace with: # # debug_cmd='eval echo "${FUNCNAME[0]} $*" >&2' bash your-script-name debug_cmd=${debug_cmd-":"} exit_cmd=: # By convention, finish your script with: # # exit $exit_status # # so that you can set exit_status to non-zero if you want to indicate # something went wrong during execution without actually bailing out at # the point of failure. exit_status=$EXIT_SUCCESS # Work around backward compatibility issue on IRIX 6.5. On IRIX 6.4+, sh # is ksh but when the shell is invoked as "sh" and the current value of # the _XPG environment variable is not equal to 1 (one), the special # positional parameter $0, within a function call, is the name of the # function. progpath=$0 # The name of this program. progname=`$ECHO "$progpath" |$SED "$sed_basename"` # Make sure we have an absolute progpath for reexecution: case $progpath in [\\/]*|[A-Za-z]:\\*) ;; *[\\/]*) progdir=`$ECHO "$progpath" |$SED "$sed_dirname"` progdir=`cd "$progdir" && pwd` progpath=$progdir/$progname ;; *) _G_IFS=$IFS IFS=${PATH_SEPARATOR-:} for progdir in $PATH; do IFS=$_G_IFS test -x "$progdir/$progname" && break done IFS=$_G_IFS test -n "$progdir" || progdir=`pwd` progpath=$progdir/$progname ;; esac ## ----------------- ## ## Standard options. ## ## ----------------- ## # The following options affect the operation of the functions defined # below, and should be set appropriately depending on run-time para- # meters passed on the command line. opt_dry_run=false opt_quiet=false opt_verbose=false # Categories 'all' and 'none' are always available. Append any others # you will pass as the first argument to func_warning from your own # code. warning_categories= # By default, display warnings according to 'opt_warning_types'. Set # 'warning_func' to ':' to elide all warnings, or func_fatal_error to # treat the next displayed warning as a fatal error. warning_func=func_warn_and_continue # Set to 'all' to display all warnings, 'none' to suppress all # warnings, or a space delimited list of some subset of # 'warning_categories' to display only the listed warnings. opt_warning_types=all ## -------------------- ## ## Resource management. ## ## -------------------- ## # This section contains definitions for functions that each ensure a # particular resource (a file, or a non-empty configuration variable for # example) is available, and if appropriate to extract default values # from pertinent package files. Call them using their associated # 'require_*' variable to ensure that they are executed, at most, once. # # It's entirely deliberate that calling these functions can set # variables that don't obey the namespace limitations obeyed by the rest # of this file, in order that that they be as useful as possible to # callers. # require_term_colors # ------------------- # Allow display of bold text on terminals that support it. require_term_colors=func_require_term_colors func_require_term_colors () { $debug_cmd test -t 1 && { # COLORTERM and USE_ANSI_COLORS environment variables take # precedence, because most terminfo databases neglect to describe # whether color sequences are supported. test -n "${COLORTERM+set}" && : ${USE_ANSI_COLORS="1"} if test 1 = "$USE_ANSI_COLORS"; then # Standard ANSI escape sequences tc_reset='' tc_bold=''; tc_standout='' tc_red=''; tc_green='' tc_blue=''; tc_cyan='' else # Otherwise trust the terminfo database after all. test -n "`tput sgr0 2>/dev/null`" && { tc_reset=`tput sgr0` test -n "`tput bold 2>/dev/null`" && tc_bold=`tput bold` tc_standout=$tc_bold test -n "`tput smso 2>/dev/null`" && tc_standout=`tput smso` test -n "`tput setaf 1 2>/dev/null`" && tc_red=`tput setaf 1` test -n "`tput setaf 2 2>/dev/null`" && tc_green=`tput setaf 2` test -n "`tput setaf 4 2>/dev/null`" && tc_blue=`tput setaf 4` test -n "`tput setaf 5 2>/dev/null`" && tc_cyan=`tput setaf 5` } fi } require_term_colors=: } ## ----------------- ## ## Function library. ## ## ----------------- ## # This section contains a variety of useful functions to call in your # scripts. Take note of the portable wrappers for features provided by # some modern shells, which will fall back to slower equivalents on # less featureful shells. # func_append VAR VALUE # --------------------- # Append VALUE onto the existing contents of VAR. # We should try to minimise forks, especially on Windows where they are # unreasonably slow, so skip the feature probes when bash or zsh are # being used: if test set = "${BASH_VERSION+set}${ZSH_VERSION+set}"; then : ${_G_HAVE_ARITH_OP="yes"} : ${_G_HAVE_XSI_OPS="yes"} # The += operator was introduced in bash 3.1 case $BASH_VERSION in [12].* | 3.0 | 3.0*) ;; *) : ${_G_HAVE_PLUSEQ_OP="yes"} ;; esac fi # _G_HAVE_PLUSEQ_OP # Can be empty, in which case the shell is probed, "yes" if += is # useable or anything else if it does not work. test -z "$_G_HAVE_PLUSEQ_OP" \ && (eval 'x=a; x+=" b"; test "a b" = "$x"') 2>/dev/null \ && _G_HAVE_PLUSEQ_OP=yes if test yes = "$_G_HAVE_PLUSEQ_OP" then # This is an XSI compatible shell, allowing a faster implementation... eval 'func_append () { $debug_cmd eval "$1+=\$2" }' else # ...otherwise fall back to using expr, which is often a shell builtin. func_append () { $debug_cmd eval "$1=\$$1\$2" } fi # func_append_quoted VAR VALUE # ---------------------------- # Quote VALUE and append to the end of shell variable VAR, separated # by a space. if test yes = "$_G_HAVE_PLUSEQ_OP"; then eval 'func_append_quoted () { $debug_cmd func_quote_for_eval "$2" eval "$1+=\\ \$func_quote_for_eval_result" }' else func_append_quoted () { $debug_cmd func_quote_for_eval "$2" eval "$1=\$$1\\ \$func_quote_for_eval_result" } fi # func_append_uniq VAR VALUE # -------------------------- # Append unique VALUE onto the existing contents of VAR, assuming # entries are delimited by the first character of VALUE. For example: # # func_append_uniq options " --another-option option-argument" # # will only append to $options if " --another-option option-argument " # is not already present somewhere in $options already (note spaces at # each end implied by leading space in second argument). func_append_uniq () { $debug_cmd eval _G_current_value='`$ECHO $'$1'`' _G_delim=`expr "$2" : '\(.\)'` case $_G_delim$_G_current_value$_G_delim in *"$2$_G_delim"*) ;; *) func_append "$@" ;; esac } # func_arith TERM... # ------------------ # Set func_arith_result to the result of evaluating TERMs. test -z "$_G_HAVE_ARITH_OP" \ && (eval 'test 2 = $(( 1 + 1 ))') 2>/dev/null \ && _G_HAVE_ARITH_OP=yes if test yes = "$_G_HAVE_ARITH_OP"; then eval 'func_arith () { $debug_cmd func_arith_result=$(( $* )) }' else func_arith () { $debug_cmd func_arith_result=`expr "$@"` } fi # func_basename FILE # ------------------ # Set func_basename_result to FILE with everything up to and including # the last / stripped. if test yes = "$_G_HAVE_XSI_OPS"; then # If this shell supports suffix pattern removal, then use it to avoid # forking. Hide the definitions single quotes in case the shell chokes # on unsupported syntax... _b='func_basename_result=${1##*/}' _d='case $1 in */*) func_dirname_result=${1%/*}$2 ;; * ) func_dirname_result=$3 ;; esac' else # ...otherwise fall back to using sed. _b='func_basename_result=`$ECHO "$1" |$SED "$sed_basename"`' _d='func_dirname_result=`$ECHO "$1" |$SED "$sed_dirname"` if test "X$func_dirname_result" = "X$1"; then func_dirname_result=$3 else func_append func_dirname_result "$2" fi' fi eval 'func_basename () { $debug_cmd '"$_b"' }' # func_dirname FILE APPEND NONDIR_REPLACEMENT # ------------------------------------------- # Compute the dirname of FILE. If nonempty, add APPEND to the result, # otherwise set result to NONDIR_REPLACEMENT. eval 'func_dirname () { $debug_cmd '"$_d"' }' # func_dirname_and_basename FILE APPEND NONDIR_REPLACEMENT # -------------------------------------------------------- # Perform func_basename and func_dirname in a single function # call: # dirname: Compute the dirname of FILE. If nonempty, # add APPEND to the result, otherwise set result # to NONDIR_REPLACEMENT. # value returned in "$func_dirname_result" # basename: Compute filename of FILE. # value retuned in "$func_basename_result" # For efficiency, we do not delegate to the functions above but instead # duplicate the functionality here. eval 'func_dirname_and_basename () { $debug_cmd '"$_b"' '"$_d"' }' # func_echo ARG... # ---------------- # Echo program name prefixed message. func_echo () { $debug_cmd _G_message=$* func_echo_IFS=$IFS IFS=$nl for _G_line in $_G_message; do IFS=$func_echo_IFS $ECHO "$progname: $_G_line" done IFS=$func_echo_IFS } # func_echo_all ARG... # -------------------- # Invoke $ECHO with all args, space-separated. func_echo_all () { $ECHO "$*" } # func_echo_infix_1 INFIX ARG... # ------------------------------ # Echo program name, followed by INFIX on the first line, with any # additional lines not showing INFIX. func_echo_infix_1 () { $debug_cmd $require_term_colors _G_infix=$1; shift _G_indent=$_G_infix _G_prefix="$progname: $_G_infix: " _G_message=$* # Strip color escape sequences before counting printable length for _G_tc in "$tc_reset" "$tc_bold" "$tc_standout" "$tc_red" "$tc_green" "$tc_blue" "$tc_cyan" do test -n "$_G_tc" && { _G_esc_tc=`$ECHO "$_G_tc" | $SED "$sed_make_literal_regex"` _G_indent=`$ECHO "$_G_indent" | $SED "s|$_G_esc_tc||g"` } done _G_indent="$progname: "`echo "$_G_indent" | $SED 's|.| |g'`" " ## exclude from sc_prohibit_nested_quotes func_echo_infix_1_IFS=$IFS IFS=$nl for _G_line in $_G_message; do IFS=$func_echo_infix_1_IFS $ECHO "$_G_prefix$tc_bold$_G_line$tc_reset" >&2 _G_prefix=$_G_indent done IFS=$func_echo_infix_1_IFS } # func_error ARG... # ----------------- # Echo program name prefixed message to standard error. func_error () { $debug_cmd $require_term_colors func_echo_infix_1 " $tc_standout${tc_red}error$tc_reset" "$*" >&2 } # func_fatal_error ARG... # ----------------------- # Echo program name prefixed message to standard error, and exit. func_fatal_error () { $debug_cmd func_error "$*" exit $EXIT_FAILURE } # func_grep EXPRESSION FILENAME # ----------------------------- # Check whether EXPRESSION matches any line of FILENAME, without output. func_grep () { $debug_cmd $GREP "$1" "$2" >/dev/null 2>&1 } # func_len STRING # --------------- # Set func_len_result to the length of STRING. STRING may not # start with a hyphen. test -z "$_G_HAVE_XSI_OPS" \ && (eval 'x=a/b/c; test 5aa/bb/cc = "${#x}${x%%/*}${x%/*}${x#*/}${x##*/}"') 2>/dev/null \ && _G_HAVE_XSI_OPS=yes if test yes = "$_G_HAVE_XSI_OPS"; then eval 'func_len () { $debug_cmd func_len_result=${#1} }' else func_len () { $debug_cmd func_len_result=`expr "$1" : ".*" 2>/dev/null || echo $max_cmd_len` } fi # func_mkdir_p DIRECTORY-PATH # --------------------------- # Make sure the entire path to DIRECTORY-PATH is available. func_mkdir_p () { $debug_cmd _G_directory_path=$1 _G_dir_list= if test -n "$_G_directory_path" && test : != "$opt_dry_run"; then # Protect directory names starting with '-' case $_G_directory_path in -*) _G_directory_path=./$_G_directory_path ;; esac # While some portion of DIR does not yet exist... while test ! -d "$_G_directory_path"; do # ...make a list in topmost first order. Use a colon delimited # list incase some portion of path contains whitespace. _G_dir_list=$_G_directory_path:$_G_dir_list # If the last portion added has no slash in it, the list is done case $_G_directory_path in */*) ;; *) break ;; esac # ...otherwise throw away the child directory and loop _G_directory_path=`$ECHO "$_G_directory_path" | $SED -e "$sed_dirname"` done _G_dir_list=`$ECHO "$_G_dir_list" | $SED 's|:*$||'` func_mkdir_p_IFS=$IFS; IFS=: for _G_dir in $_G_dir_list; do IFS=$func_mkdir_p_IFS # mkdir can fail with a 'File exist' error if two processes # try to create one of the directories concurrently. Don't # stop in that case! $MKDIR "$_G_dir" 2>/dev/null || : done IFS=$func_mkdir_p_IFS # Bail out if we (or some other process) failed to create a directory. test -d "$_G_directory_path" || \ func_fatal_error "Failed to create '$1'" fi } # func_mktempdir [BASENAME] # ------------------------- # Make a temporary directory that won't clash with other running # libtool processes, and avoids race conditions if possible. If # given, BASENAME is the basename for that directory. func_mktempdir () { $debug_cmd _G_template=${TMPDIR-/tmp}/${1-$progname} if test : = "$opt_dry_run"; then # Return a directory name, but don't create it in dry-run mode _G_tmpdir=$_G_template-$$ else # If mktemp works, use that first and foremost _G_tmpdir=`mktemp -d "$_G_template-XXXXXXXX" 2>/dev/null` if test ! -d "$_G_tmpdir"; then # Failing that, at least try and use $RANDOM to avoid a race _G_tmpdir=$_G_template-${RANDOM-0}$$ func_mktempdir_umask=`umask` umask 0077 $MKDIR "$_G_tmpdir" umask $func_mktempdir_umask fi # If we're not in dry-run mode, bomb out on failure test -d "$_G_tmpdir" || \ func_fatal_error "cannot create temporary directory '$_G_tmpdir'" fi $ECHO "$_G_tmpdir" } # func_normal_abspath PATH # ------------------------ # Remove doubled-up and trailing slashes, "." path components, # and cancel out any ".." path components in PATH after making # it an absolute path. func_normal_abspath () { $debug_cmd # These SED scripts presuppose an absolute path with a trailing slash. _G_pathcar='s|^/\([^/]*\).*$|\1|' _G_pathcdr='s|^/[^/]*||' _G_removedotparts=':dotsl s|/\./|/|g t dotsl s|/\.$|/|' _G_collapseslashes='s|/\{1,\}|/|g' _G_finalslash='s|/*$|/|' # Start from root dir and reassemble the path. func_normal_abspath_result= func_normal_abspath_tpath=$1 func_normal_abspath_altnamespace= case $func_normal_abspath_tpath in "") # Empty path, that just means $cwd. func_stripname '' '/' "`pwd`" func_normal_abspath_result=$func_stripname_result return ;; # The next three entries are used to spot a run of precisely # two leading slashes without using negated character classes; # we take advantage of case's first-match behaviour. ///*) # Unusual form of absolute path, do nothing. ;; //*) # Not necessarily an ordinary path; POSIX reserves leading '//' # and for example Cygwin uses it to access remote file shares # over CIFS/SMB, so we conserve a leading double slash if found. func_normal_abspath_altnamespace=/ ;; /*) # Absolute path, do nothing. ;; *) # Relative path, prepend $cwd. func_normal_abspath_tpath=`pwd`/$func_normal_abspath_tpath ;; esac # Cancel out all the simple stuff to save iterations. We also want # the path to end with a slash for ease of parsing, so make sure # there is one (and only one) here. func_normal_abspath_tpath=`$ECHO "$func_normal_abspath_tpath" | $SED \ -e "$_G_removedotparts" -e "$_G_collapseslashes" -e "$_G_finalslash"` while :; do # Processed it all yet? if test / = "$func_normal_abspath_tpath"; then # If we ascended to the root using ".." the result may be empty now. if test -z "$func_normal_abspath_result"; then func_normal_abspath_result=/ fi break fi func_normal_abspath_tcomponent=`$ECHO "$func_normal_abspath_tpath" | $SED \ -e "$_G_pathcar"` func_normal_abspath_tpath=`$ECHO "$func_normal_abspath_tpath" | $SED \ -e "$_G_pathcdr"` # Figure out what to do with it case $func_normal_abspath_tcomponent in "") # Trailing empty path component, ignore it. ;; ..) # Parent dir; strip last assembled component from result. func_dirname "$func_normal_abspath_result" func_normal_abspath_result=$func_dirname_result ;; *) # Actual path component, append it. func_append func_normal_abspath_result "/$func_normal_abspath_tcomponent" ;; esac done # Restore leading double-slash if one was found on entry. func_normal_abspath_result=$func_normal_abspath_altnamespace$func_normal_abspath_result } # func_notquiet ARG... # -------------------- # Echo program name prefixed message only when not in quiet mode. func_notquiet () { $debug_cmd $opt_quiet || func_echo ${1+"$@"} # A bug in bash halts the script if the last line of a function # fails when set -e is in force, so we need another command to # work around that: : } # func_relative_path SRCDIR DSTDIR # -------------------------------- # Set func_relative_path_result to the relative path from SRCDIR to DSTDIR. func_relative_path () { $debug_cmd func_relative_path_result= func_normal_abspath "$1" func_relative_path_tlibdir=$func_normal_abspath_result func_normal_abspath "$2" func_relative_path_tbindir=$func_normal_abspath_result # Ascend the tree starting from libdir while :; do # check if we have found a prefix of bindir case $func_relative_path_tbindir in $func_relative_path_tlibdir) # found an exact match func_relative_path_tcancelled= break ;; $func_relative_path_tlibdir*) # found a matching prefix func_stripname "$func_relative_path_tlibdir" '' "$func_relative_path_tbindir" func_relative_path_tcancelled=$func_stripname_result if test -z "$func_relative_path_result"; then func_relative_path_result=. fi break ;; *) func_dirname $func_relative_path_tlibdir func_relative_path_tlibdir=$func_dirname_result if test -z "$func_relative_path_tlibdir"; then # Have to descend all the way to the root! func_relative_path_result=../$func_relative_path_result func_relative_path_tcancelled=$func_relative_path_tbindir break fi func_relative_path_result=../$func_relative_path_result ;; esac done # Now calculate path; take care to avoid doubling-up slashes. func_stripname '' '/' "$func_relative_path_result" func_relative_path_result=$func_stripname_result func_stripname '/' '/' "$func_relative_path_tcancelled" if test -n "$func_stripname_result"; then func_append func_relative_path_result "/$func_stripname_result" fi # Normalisation. If bindir is libdir, return '.' else relative path. if test -n "$func_relative_path_result"; then func_stripname './' '' "$func_relative_path_result" func_relative_path_result=$func_stripname_result fi test -n "$func_relative_path_result" || func_relative_path_result=. : } # func_quote_for_eval ARG... # -------------------------- # Aesthetically quote ARGs to be evaled later. # This function returns two values: # i) func_quote_for_eval_result # double-quoted, suitable for a subsequent eval # ii) func_quote_for_eval_unquoted_result # has all characters that are still active within double # quotes backslashified. func_quote_for_eval () { $debug_cmd func_quote_for_eval_unquoted_result= func_quote_for_eval_result= while test 0 -lt $#; do case $1 in *[\\\`\"\$]*) _G_unquoted_arg=`printf '%s\n' "$1" |$SED "$sed_quote_subst"` ;; *) _G_unquoted_arg=$1 ;; esac if test -n "$func_quote_for_eval_unquoted_result"; then func_append func_quote_for_eval_unquoted_result " $_G_unquoted_arg" else func_append func_quote_for_eval_unquoted_result "$_G_unquoted_arg" fi case $_G_unquoted_arg in # Double-quote args containing shell metacharacters to delay # word splitting, command substitution and variable expansion # for a subsequent eval. # Many Bourne shells cannot handle close brackets correctly # in scan sets, so we specify it separately. *[\[\~\#\^\&\*\(\)\{\}\|\;\<\>\?\'\ \ ]*|*]*|"") _G_quoted_arg=\"$_G_unquoted_arg\" ;; *) _G_quoted_arg=$_G_unquoted_arg ;; esac if test -n "$func_quote_for_eval_result"; then func_append func_quote_for_eval_result " $_G_quoted_arg" else func_append func_quote_for_eval_result "$_G_quoted_arg" fi shift done } # func_quote_for_expand ARG # ------------------------- # Aesthetically quote ARG to be evaled later; same as above, # but do not quote variable references. func_quote_for_expand () { $debug_cmd case $1 in *[\\\`\"]*) _G_arg=`$ECHO "$1" | $SED \ -e "$sed_double_quote_subst" -e "$sed_double_backslash"` ;; *) _G_arg=$1 ;; esac case $_G_arg in # Double-quote args containing shell metacharacters to delay # word splitting and command substitution for a subsequent eval. # Many Bourne shells cannot handle close brackets correctly # in scan sets, so we specify it separately. *[\[\~\#\^\&\*\(\)\{\}\|\;\<\>\?\'\ \ ]*|*]*|"") _G_arg=\"$_G_arg\" ;; esac func_quote_for_expand_result=$_G_arg } # func_stripname PREFIX SUFFIX NAME # --------------------------------- # strip PREFIX and SUFFIX from NAME, and store in func_stripname_result. # PREFIX and SUFFIX must not contain globbing or regex special # characters, hashes, percent signs, but SUFFIX may contain a leading # dot (in which case that matches only a dot). if test yes = "$_G_HAVE_XSI_OPS"; then eval 'func_stripname () { $debug_cmd # pdksh 5.2.14 does not do ${X%$Y} correctly if both X and Y are # positional parameters, so assign one to ordinary variable first. func_stripname_result=$3 func_stripname_result=${func_stripname_result#"$1"} func_stripname_result=${func_stripname_result%"$2"} }' else func_stripname () { $debug_cmd case $2 in .*) func_stripname_result=`$ECHO "$3" | $SED -e "s%^$1%%" -e "s%\\\\$2\$%%"`;; *) func_stripname_result=`$ECHO "$3" | $SED -e "s%^$1%%" -e "s%$2\$%%"`;; esac } fi # func_show_eval CMD [FAIL_EXP] # ----------------------------- # Unless opt_quiet is true, then output CMD. Then, if opt_dryrun is # not true, evaluate CMD. If the evaluation of CMD fails, and FAIL_EXP # is given, then evaluate it. func_show_eval () { $debug_cmd _G_cmd=$1 _G_fail_exp=${2-':'} func_quote_for_expand "$_G_cmd" eval "func_notquiet $func_quote_for_expand_result" $opt_dry_run || { eval "$_G_cmd" _G_status=$? if test 0 -ne "$_G_status"; then eval "(exit $_G_status); $_G_fail_exp" fi } } # func_show_eval_locale CMD [FAIL_EXP] # ------------------------------------ # Unless opt_quiet is true, then output CMD. Then, if opt_dryrun is # not true, evaluate CMD. If the evaluation of CMD fails, and FAIL_EXP # is given, then evaluate it. Use the saved locale for evaluation. func_show_eval_locale () { $debug_cmd _G_cmd=$1 _G_fail_exp=${2-':'} $opt_quiet || { func_quote_for_expand "$_G_cmd" eval "func_echo $func_quote_for_expand_result" } $opt_dry_run || { eval "$_G_user_locale $_G_cmd" _G_status=$? eval "$_G_safe_locale" if test 0 -ne "$_G_status"; then eval "(exit $_G_status); $_G_fail_exp" fi } } # func_tr_sh # ---------- # Turn $1 into a string suitable for a shell variable name. # Result is stored in $func_tr_sh_result. All characters # not in the set a-zA-Z0-9_ are replaced with '_'. Further, # if $1 begins with a digit, a '_' is prepended as well. func_tr_sh () { $debug_cmd case $1 in [0-9]* | *[!a-zA-Z0-9_]*) func_tr_sh_result=`$ECHO "$1" | $SED -e 's/^\([0-9]\)/_\1/' -e 's/[^a-zA-Z0-9_]/_/g'` ;; * ) func_tr_sh_result=$1 ;; esac } # func_verbose ARG... # ------------------- # Echo program name prefixed message in verbose mode only. func_verbose () { $debug_cmd $opt_verbose && func_echo "$*" : } # func_warn_and_continue ARG... # ----------------------------- # Echo program name prefixed warning message to standard error. func_warn_and_continue () { $debug_cmd $require_term_colors func_echo_infix_1 "${tc_red}warning$tc_reset" "$*" >&2 } # func_warning CATEGORY ARG... # ---------------------------- # Echo program name prefixed warning message to standard error. Warning # messages can be filtered according to CATEGORY, where this function # elides messages where CATEGORY is not listed in the global variable # 'opt_warning_types'. func_warning () { $debug_cmd # CATEGORY must be in the warning_categories list! case " $warning_categories " in *" $1 "*) ;; *) func_internal_error "invalid warning category '$1'" ;; esac _G_category=$1 shift case " $opt_warning_types " in *" $_G_category "*) $warning_func ${1+"$@"} ;; esac } # func_sort_ver VER1 VER2 # ----------------------- # 'sort -V' is not generally available. # Note this deviates from the version comparison in automake # in that it treats 1.5 < 1.5.0, and treats 1.4.4a < 1.4-p3a # but this should suffice as we won't be specifying old # version formats or redundant trailing .0 in bootstrap.conf. # If we did want full compatibility then we should probably # use m4_version_compare from autoconf. func_sort_ver () { $debug_cmd printf '%s\n%s\n' "$1" "$2" \ | sort -t. -k 1,1n -k 2,2n -k 3,3n -k 4,4n -k 5,5n -k 6,6n -k 7,7n -k 8,8n -k 9,9n } # func_lt_ver PREV CURR # --------------------- # Return true if PREV and CURR are in the correct order according to # func_sort_ver, otherwise false. Use it like this: # # func_lt_ver "$prev_ver" "$proposed_ver" || func_fatal_error "..." func_lt_ver () { $debug_cmd test "x$1" = x`func_sort_ver "$1" "$2" | $SED 1q` } # Local variables: # mode: shell-script # sh-indentation: 2 # eval: (add-hook 'before-save-hook 'time-stamp) # time-stamp-pattern: "10/scriptversion=%:y-%02m-%02d.%02H; # UTC" # time-stamp-time-zone: "UTC" # End: #! /bin/sh # Set a version string for this script. scriptversion=2014-01-07.03; # UTC # A portable, pluggable option parser for Bourne shell. # Written by Gary V. Vaughan, 2010 # Copyright (C) 2010-2015 Free Software Foundation, Inc. # This is free software; see the source for copying conditions. There is NO # warranty; not even for MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. # This program is free software: you can redistribute it and/or modify # it under the terms of the GNU General Public License as published by # the Free Software Foundation, either version 3 of the License, or # (at your option) any later version. # This program is distributed in the hope that it will be useful, # but WITHOUT ANY WARRANTY; without even the implied warranty of # MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the # GNU General Public License for more details. # You should have received a copy of the GNU General Public License # along with this program. If not, see <http://www.gnu.org/licenses/>. # Please report bugs or propose patches to gary@gnu.org. ## ------ ## ## Usage. ## ## ------ ## # This file is a library for parsing options in your shell scripts along # with assorted other useful supporting features that you can make use # of too. # # For the simplest scripts you might need only: # # #!/bin/sh # . relative/path/to/funclib.sh # . relative/path/to/options-parser # scriptversion=1.0 # func_options ${1+"$@"} # eval set dummy "$func_options_result"; shift # ...rest of your script... # # In order for the '--version' option to work, you will need to have a # suitably formatted comment like the one at the top of this file # starting with '# Written by ' and ending with '# warranty; '. # # For '-h' and '--help' to work, you will also need a one line # description of your script's purpose in a comment directly above the # '# Written by ' line, like the one at the top of this file. # # The default options also support '--debug', which will turn on shell # execution tracing (see the comment above debug_cmd below for another # use), and '--verbose' and the func_verbose function to allow your script # to display verbose messages only when your user has specified # '--verbose'. # # After sourcing this file, you can plug processing for additional # options by amending the variables from the 'Configuration' section # below, and following the instructions in the 'Option parsing' # section further down. ## -------------- ## ## Configuration. ## ## -------------- ## # You should override these variables in your script after sourcing this # file so that they reflect the customisations you have added to the # option parser. # The usage line for option parsing errors and the start of '-h' and # '--help' output messages. You can embed shell variables for delayed # expansion at the time the message is displayed, but you will need to # quote other shell meta-characters carefully to prevent them being # expanded when the contents are evaled. usage='$progpath [OPTION]...' # Short help message in response to '-h' and '--help'. Add to this or # override it after sourcing this library to reflect the full set of # options your script accepts. usage_message="\ --debug enable verbose shell tracing -W, --warnings=CATEGORY report the warnings falling in CATEGORY [all] -v, --verbose verbosely report processing --version print version information and exit -h, --help print short or long help message and exit " # Additional text appended to 'usage_message' in response to '--help'. long_help_message=" Warning categories include: 'all' show all warnings 'none' turn off all the warnings 'error' warnings are treated as fatal errors" # Help message printed before fatal option parsing errors. fatal_help="Try '\$progname --help' for more information." ## ------------------------- ## ## Hook function management. ## ## ------------------------- ## # This section contains functions for adding, removing, and running hooks # to the main code. A hook is just a named list of of function, that can # be run in order later on. # func_hookable FUNC_NAME # ----------------------- # Declare that FUNC_NAME will run hooks added with # 'func_add_hook FUNC_NAME ...'. func_hookable () { $debug_cmd func_append hookable_fns " $1" } # func_add_hook FUNC_NAME HOOK_FUNC # --------------------------------- # Request that FUNC_NAME call HOOK_FUNC before it returns. FUNC_NAME must # first have been declared "hookable" by a call to 'func_hookable'. func_add_hook () { $debug_cmd case " $hookable_fns " in *" $1 "*) ;; *) func_fatal_error "'$1' does not accept hook functions." ;; esac eval func_append ${1}_hooks '" $2"' } # func_remove_hook FUNC_NAME HOOK_FUNC # ------------------------------------ # Remove HOOK_FUNC from the list of functions called by FUNC_NAME. func_remove_hook () { $debug_cmd eval ${1}_hooks='`$ECHO "\$'$1'_hooks" |$SED "s| '$2'||"`' } # func_run_hooks FUNC_NAME [ARG]... # --------------------------------- # Run all hook functions registered to FUNC_NAME. # It is assumed that the list of hook functions contains nothing more # than a whitespace-delimited list of legal shell function names, and # no effort is wasted trying to catch shell meta-characters or preserve # whitespace. func_run_hooks () { $debug_cmd case " $hookable_fns " in *" $1 "*) ;; *) func_fatal_error "'$1' does not support hook funcions.n" ;; esac eval _G_hook_fns=\$$1_hooks; shift for _G_hook in $_G_hook_fns; do eval $_G_hook '"$@"' # store returned options list back into positional # parameters for next 'cmd' execution. eval _G_hook_result=\$${_G_hook}_result eval set dummy "$_G_hook_result"; shift done func_quote_for_eval ${1+"$@"} func_run_hooks_result=$func_quote_for_eval_result } ## --------------- ## ## Option parsing. ## ## --------------- ## # In order to add your own option parsing hooks, you must accept the # full positional parameter list in your hook function, remove any # options that you action, and then pass back the remaining unprocessed # options in '<hooked_function_name>_result', escaped suitably for # 'eval'. Like this: # # my_options_prep () # { # $debug_cmd # # # Extend the existing usage message. # usage_message=$usage_message' # -s, --silent don'\''t print informational messages # ' # # func_quote_for_eval ${1+"$@"} # my_options_prep_result=$func_quote_for_eval_result # } # func_add_hook func_options_prep my_options_prep # # # my_silent_option () # { # $debug_cmd # # # Note that for efficiency, we parse as many options as we can # # recognise in a loop before passing the remainder back to the # # caller on the first unrecognised argument we encounter. # while test $# -gt 0; do # opt=$1; shift # case $opt in # --silent|-s) opt_silent=: ;; # # Separate non-argument short options: # -s*) func_split_short_opt "$_G_opt" # set dummy "$func_split_short_opt_name" \ # "-$func_split_short_opt_arg" ${1+"$@"} # shift # ;; # *) set dummy "$_G_opt" "$*"; shift; break ;; # esac # done # # func_quote_for_eval ${1+"$@"} # my_silent_option_result=$func_quote_for_eval_result # } # func_add_hook func_parse_options my_silent_option # # # my_option_validation () # { # $debug_cmd # # $opt_silent && $opt_verbose && func_fatal_help "\ # '--silent' and '--verbose' options are mutually exclusive." # # func_quote_for_eval ${1+"$@"} # my_option_validation_result=$func_quote_for_eval_result # } # func_add_hook func_validate_options my_option_validation # # You'll alse need to manually amend $usage_message to reflect the extra # options you parse. It's preferable to append if you can, so that # multiple option parsing hooks can be added safely. # func_options [ARG]... # --------------------- # All the functions called inside func_options are hookable. See the # individual implementations for details. func_hookable func_options func_options () { $debug_cmd func_options_prep ${1+"$@"} eval func_parse_options \ ${func_options_prep_result+"$func_options_prep_result"} eval func_validate_options \ ${func_parse_options_result+"$func_parse_options_result"} eval func_run_hooks func_options \ ${func_validate_options_result+"$func_validate_options_result"} # save modified positional parameters for caller func_options_result=$func_run_hooks_result } # func_options_prep [ARG]... # -------------------------- # All initialisations required before starting the option parse loop. # Note that when calling hook functions, we pass through the list of # positional parameters. If a hook function modifies that list, and # needs to propogate that back to rest of this script, then the complete # modified list must be put in 'func_run_hooks_result' before # returning. func_hookable func_options_prep func_options_prep () { $debug_cmd # Option defaults: opt_verbose=false opt_warning_types= func_run_hooks func_options_prep ${1+"$@"} # save modified positional parameters for caller func_options_prep_result=$func_run_hooks_result } # func_parse_options [ARG]... # --------------------------- # The main option parsing loop. func_hookable func_parse_options func_parse_options () { $debug_cmd func_parse_options_result= # this just eases exit handling while test $# -gt 0; do # Defer to hook functions for initial option parsing, so they # get priority in the event of reusing an option name. func_run_hooks func_parse_options ${1+"$@"} # Adjust func_parse_options positional parameters to match eval set dummy "$func_run_hooks_result"; shift # Break out of the loop if we already parsed every option. test $# -gt 0 || break _G_opt=$1 shift case $_G_opt in --debug|-x) debug_cmd='set -x' func_echo "enabling shell trace mode" $debug_cmd ;; --no-warnings|--no-warning|--no-warn) set dummy --warnings none ${1+"$@"} shift ;; --warnings|--warning|-W) test $# = 0 && func_missing_arg $_G_opt && break case " $warning_categories $1" in *" $1 "*) # trailing space prevents matching last $1 above func_append_uniq opt_warning_types " $1" ;; *all) opt_warning_types=$warning_categories ;; *none) opt_warning_types=none warning_func=: ;; *error) opt_warning_types=$warning_categories warning_func=func_fatal_error ;; *) func_fatal_error \ "unsupported warning category: '$1'" ;; esac shift ;; --verbose|-v) opt_verbose=: ;; --version) func_version ;; -\?|-h) func_usage ;; --help) func_help ;; # Separate optargs to long options (plugins may need this): --*=*) func_split_equals "$_G_opt" set dummy "$func_split_equals_lhs" \ "$func_split_equals_rhs" ${1+"$@"} shift ;; # Separate optargs to short options: -W*) func_split_short_opt "$_G_opt" set dummy "$func_split_short_opt_name" \ "$func_split_short_opt_arg" ${1+"$@"} shift ;; # Separate non-argument short options: -\?*|-h*|-v*|-x*) func_split_short_opt "$_G_opt" set dummy "$func_split_short_opt_name" \ "-$func_split_short_opt_arg" ${1+"$@"} shift ;; --) break ;; -*) func_fatal_help "unrecognised option: '$_G_opt'" ;; *) set dummy "$_G_opt" ${1+"$@"}; shift; break ;; esac done # save modified positional parameters for caller func_quote_for_eval ${1+"$@"} func_parse_options_result=$func_quote_for_eval_result } # func_validate_options [ARG]... # ------------------------------ # Perform any sanity checks on option settings and/or unconsumed # arguments. func_hookable func_validate_options func_validate_options () { $debug_cmd # Display all warnings if -W was not given. test -n "$opt_warning_types" || opt_warning_types=" $warning_categories" func_run_hooks func_validate_options ${1+"$@"} # Bail if the options were screwed! $exit_cmd $EXIT_FAILURE # save modified positional parameters for caller func_validate_options_result=$func_run_hooks_result } ## ----------------- ## ## Helper functions. ## ## ----------------- ## # This section contains the helper functions used by the rest of the # hookable option parser framework in ascii-betical order. # func_fatal_help ARG... # ---------------------- # Echo program name prefixed message to standard error, followed by # a help hint, and exit. func_fatal_help () { $debug_cmd eval \$ECHO \""Usage: $usage"\" eval \$ECHO \""$fatal_help"\" func_error ${1+"$@"} exit $EXIT_FAILURE } # func_help # --------- # Echo long help message to standard output and exit. func_help () { $debug_cmd func_usage_message $ECHO "$long_help_message" exit 0 } # func_missing_arg ARGNAME # ------------------------ # Echo program name prefixed message to standard error and set global # exit_cmd. func_missing_arg () { $debug_cmd func_error "Missing argument for '$1'." exit_cmd=exit } # func_split_equals STRING # ------------------------ # Set func_split_equals_lhs and func_split_equals_rhs shell variables after # splitting STRING at the '=' sign. test -z "$_G_HAVE_XSI_OPS" \ && (eval 'x=a/b/c; test 5aa/bb/cc = "${#x}${x%%/*}${x%/*}${x#*/}${x##*/}"') 2>/dev/null \ && _G_HAVE_XSI_OPS=yes if test yes = "$_G_HAVE_XSI_OPS" then # This is an XSI compatible shell, allowing a faster implementation... eval 'func_split_equals () { $debug_cmd func_split_equals_lhs=${1%%=*} func_split_equals_rhs=${1#*=} test "x$func_split_equals_lhs" = "x$1" \ && func_split_equals_rhs= }' else # ...otherwise fall back to using expr, which is often a shell builtin. func_split_equals () { $debug_cmd func_split_equals_lhs=`expr "x$1" : 'x\([^=]*\)'` func_split_equals_rhs= test "x$func_split_equals_lhs" = "x$1" \ || func_split_equals_rhs=`expr "x$1" : 'x[^=]*=\(.*\)$'` } fi #func_split_equals # func_split_short_opt SHORTOPT # ----------------------------- # Set func_split_short_opt_name and func_split_short_opt_arg shell # variables after splitting SHORTOPT after the 2nd character. if test yes = "$_G_HAVE_XSI_OPS" then # This is an XSI compatible shell, allowing a faster implementation... eval 'func_split_short_opt () { $debug_cmd func_split_short_opt_arg=${1#??} func_split_short_opt_name=${1%"$func_split_short_opt_arg"} }' else # ...otherwise fall back to using expr, which is often a shell builtin. func_split_short_opt () { $debug_cmd func_split_short_opt_name=`expr "x$1" : 'x-\(.\)'` func_split_short_opt_arg=`expr "x$1" : 'x-.\(.*\)$'` } fi #func_split_short_opt # func_usage # ---------- # Echo short help message to standard output and exit. func_usage () { $debug_cmd func_usage_message $ECHO "Run '$progname --help |${PAGER-more}' for full usage" exit 0 } # func_usage_message # ------------------ # Echo short help message to standard output. func_usage_message () { $debug_cmd eval \$ECHO \""Usage: $usage"\" echo $SED -n 's|^# || /^Written by/{ x;p;x } h /^Written by/q' < "$progpath" echo eval \$ECHO \""$usage_message"\" } # func_version # ------------ # Echo version message to standard output and exit. func_version () { $debug_cmd printf '%s\n' "$progname $scriptversion" $SED -n ' /(C)/!b go :more /\./!{ N s|\n# | | b more } :go /^# Written by /,/# warranty; / { s|^# || s|^# *$|| s|\((C)\)[ 0-9,-]*[ ,-]\([1-9][0-9]* \)|\1 \2| p } /^# Written by / { s|^# || p } /^warranty; /q' < "$progpath" exit $? } # Local variables: # mode: shell-script # sh-indentation: 2 # eval: (add-hook 'before-save-hook 'time-stamp) # time-stamp-pattern: "10/scriptversion=%:y-%02m-%02d.%02H; # UTC" # time-stamp-time-zone: "UTC" # End: # Set a version string. scriptversion='(GNU libtool) 2.4.6' # func_echo ARG... # ---------------- # Libtool also displays the current mode in messages, so override # funclib.sh func_echo with this custom definition. func_echo () { $debug_cmd _G_message=$* func_echo_IFS=$IFS IFS=$nl for _G_line in $_G_message; do IFS=$func_echo_IFS $ECHO "$progname${opt_mode+: $opt_mode}: $_G_line" done IFS=$func_echo_IFS } # func_warning ARG... # ------------------- # Libtool warnings are not categorized, so override funclib.sh # func_warning with this simpler definition. func_warning () { $debug_cmd $warning_func ${1+"$@"} } ## ---------------- ## ## Options parsing. ## ## ---------------- ## # Hook in the functions to make sure our own options are parsed during # the option parsing loop. usage='$progpath [OPTION]... [MODE-ARG]...' # Short help message in response to '-h'. usage_message="Options: --config show all configuration variables --debug enable verbose shell tracing -n, --dry-run display commands without modifying any files --features display basic configuration information and exit --mode=MODE use operation mode MODE --no-warnings equivalent to '-Wnone' --preserve-dup-deps don't remove duplicate dependency libraries --quiet, --silent don't print informational messages --tag=TAG use configuration variables from tag TAG -v, --verbose print more informational messages than default --version print version information -W, --warnings=CATEGORY report the warnings falling in CATEGORY [all] -h, --help, --help-all print short, long, or detailed help message " # Additional text appended to 'usage_message' in response to '--help'. func_help () { $debug_cmd func_usage_message $ECHO "$long_help_message MODE must be one of the following: clean remove files from the build directory compile compile a source file into a libtool object execute automatically set library path, then run a program finish complete the installation of libtool libraries install install libraries or executables link create a library or an executable uninstall remove libraries from an installed directory MODE-ARGS vary depending on the MODE. When passed as first option, '--mode=MODE' may be abbreviated as 'MODE' or a unique abbreviation of that. Try '$progname --help --mode=MODE' for a more detailed description of MODE. When reporting a bug, please describe a test case to reproduce it and include the following information: host-triplet: $host shell: $SHELL compiler: $LTCC compiler flags: $LTCFLAGS linker: $LD (gnu? $with_gnu_ld) version: $progname $scriptversion Debian-2.4.6-2 automake: `($AUTOMAKE --version) 2>/dev/null |$SED 1q` autoconf: `($AUTOCONF --version) 2>/dev/null |$SED 1q` Report bugs to <bug-libtool@gnu.org>. GNU libtool home page: <http://www.gnu.org/s/libtool/>. General help using GNU software: <http://www.gnu.org/gethelp/>." exit 0 } # func_lo2o OBJECT-NAME # --------------------- # Transform OBJECT-NAME from a '.lo' suffix to the platform specific # object suffix. lo2o=s/\\.lo\$/.$objext/ o2lo=s/\\.$objext\$/.lo/ if test yes = "$_G_HAVE_XSI_OPS"; then eval 'func_lo2o () { case $1 in *.lo) func_lo2o_result=${1%.lo}.$objext ;; * ) func_lo2o_result=$1 ;; esac }' # func_xform LIBOBJ-OR-SOURCE # --------------------------- # Transform LIBOBJ-OR-SOURCE from a '.o' or '.c' (or otherwise) # suffix to a '.lo' libtool-object suffix. eval 'func_xform () { func_xform_result=${1%.*}.lo }' else # ...otherwise fall back to using sed. func_lo2o () { func_lo2o_result=`$ECHO "$1" | $SED "$lo2o"` } func_xform () { func_xform_result=`$ECHO "$1" | $SED 's|\.[^.]*$|.lo|'` } fi # func_fatal_configuration ARG... # ------------------------------- # Echo program name prefixed message to standard error, followed by # a configuration failure hint, and exit. func_fatal_configuration () { func__fatal_error ${1+"$@"} \ "See the $PACKAGE documentation for more information." \ "Fatal configuration error." } # func_config # ----------- # Display the configuration for all the tags in this script. func_config () { re_begincf='^# ### BEGIN LIBTOOL' re_endcf='^# ### END LIBTOOL' # Default configuration. $SED "1,/$re_begincf CONFIG/d;/$re_endcf CONFIG/,\$d" < "$progpath" # Now print the configurations for the tags. for tagname in $taglist; do $SED -n "/$re_begincf TAG CONFIG: $tagname\$/,/$re_endcf TAG CONFIG: $tagname\$/p" < "$progpath" done exit $? } # func_features # ------------- # Display the features supported by this script. func_features () { echo "host: $host" if test yes = "$build_libtool_libs"; then echo "enable shared libraries" else echo "disable shared libraries" fi if test yes = "$build_old_libs"; then echo "enable static libraries" else echo "disable static libraries" fi exit $? } # func_enable_tag TAGNAME # ----------------------- # Verify that TAGNAME is valid, and either flag an error and exit, or # enable the TAGNAME tag. We also add TAGNAME to the global $taglist # variable here. func_enable_tag () { # Global variable: tagname=$1 re_begincf="^# ### BEGIN LIBTOOL TAG CONFIG: $tagname\$" re_endcf="^# ### END LIBTOOL TAG CONFIG: $tagname\$" sed_extractcf=/$re_begincf/,/$re_endcf/p # Validate tagname. case $tagname in *[!-_A-Za-z0-9,/]*) func_fatal_error "invalid tag name: $tagname" ;; esac # Don't test for the "default" C tag, as we know it's # there but not specially marked. case $tagname in CC) ;; *) if $GREP "$re_begincf" "$progpath" >/dev/null 2>&1; then taglist="$taglist $tagname" # Evaluate the configuration. Be careful to quote the path # and the sed script, to avoid splitting on whitespace, but # also don't use non-portable quotes within backquotes within # quotes we have to do it in 2 steps: extractedcf=`$SED -n -e "$sed_extractcf" < "$progpath"` eval "$extractedcf" else func_error "ignoring unknown tag $tagname" fi ;; esac } # func_check_version_match # ------------------------ # Ensure that we are using m4 macros, and libtool script from the same # release of libtool. func_check_version_match () { if test "$package_revision" != "$macro_revision"; then if test "$VERSION" != "$macro_version"; then if test -z "$macro_version"; then cat >&2 <<_LT_EOF $progname: Version mismatch error. This is $PACKAGE $VERSION, but the $progname: definition of this LT_INIT comes from an older release. $progname: You should recreate aclocal.m4 with macros from $PACKAGE $VERSION $progname: and run autoconf again. _LT_EOF else cat >&2 <<_LT_EOF $progname: Version mismatch error. This is $PACKAGE $VERSION, but the $progname: definition of this LT_INIT comes from $PACKAGE $macro_version. $progname: You should recreate aclocal.m4 with macros from $PACKAGE $VERSION $progname: and run autoconf again. _LT_EOF fi else cat >&2 <<_LT_EOF $progname: Version mismatch error. This is $PACKAGE $VERSION, revision $package_revision, $progname: but the definition of this LT_INIT comes from revision $macro_revision. $progname: You should recreate aclocal.m4 with macros from revision $package_revision $progname: of $PACKAGE $VERSION and run autoconf again. _LT_EOF fi exit $EXIT_MISMATCH fi } # libtool_options_prep [ARG]... # ----------------------------- # Preparation for options parsed by libtool. libtool_options_prep () { $debug_mode # Option defaults: opt_config=false opt_dlopen= opt_dry_run=false opt_help=false opt_mode= opt_preserve_dup_deps=false opt_quiet=false nonopt= preserve_args= # Shorthand for --mode=foo, only valid as the first argument case $1 in clean|clea|cle|cl) shift; set dummy --mode clean ${1+"$@"}; shift ;; compile|compil|compi|comp|com|co|c) shift; set dummy --mode compile ${1+"$@"}; shift ;; execute|execut|execu|exec|exe|ex|e) shift; set dummy --mode execute ${1+"$@"}; shift ;; finish|finis|fini|fin|fi|f) shift; set dummy --mode finish ${1+"$@"}; shift ;; install|instal|insta|inst|ins|in|i) shift; set dummy --mode install ${1+"$@"}; shift ;; link|lin|li|l) shift; set dummy --mode link ${1+"$@"}; shift ;; uninstall|uninstal|uninsta|uninst|unins|unin|uni|un|u) shift; set dummy --mode uninstall ${1+"$@"}; shift ;; esac # Pass back the list of options. func_quote_for_eval ${1+"$@"} libtool_options_prep_result=$func_quote_for_eval_result } func_add_hook func_options_prep libtool_options_prep # libtool_parse_options [ARG]... # --------------------------------- # Provide handling for libtool specific options. libtool_parse_options () { $debug_cmd # Perform our own loop to consume as many options as possible in # each iteration. while test $# -gt 0; do _G_opt=$1 shift case $_G_opt in --dry-run|--dryrun|-n) opt_dry_run=: ;; --config) func_config ;; --dlopen|-dlopen) opt_dlopen="${opt_dlopen+$opt_dlopen }$1" shift ;; --preserve-dup-deps) opt_preserve_dup_deps=: ;; --features) func_features ;; --finish) set dummy --mode finish ${1+"$@"}; shift ;; --help) opt_help=: ;; --help-all) opt_help=': help-all' ;; --mode) test $# = 0 && func_missing_arg $_G_opt && break opt_mode=$1 case $1 in # Valid mode arguments: clean|compile|execute|finish|install|link|relink|uninstall) ;; # Catch anything else as an error *) func_error "invalid argument for $_G_opt" exit_cmd=exit break ;; esac shift ;; --no-silent|--no-quiet) opt_quiet=false func_append preserve_args " $_G_opt" ;; --no-warnings|--no-warning|--no-warn) opt_warning=false func_append preserve_args " $_G_opt" ;; --no-verbose) opt_verbose=false func_append preserve_args " $_G_opt" ;; --silent|--quiet) opt_quiet=: opt_verbose=false func_append preserve_args " $_G_opt" ;; --tag) test $# = 0 && func_missing_arg $_G_opt && break opt_tag=$1 func_append preserve_args " $_G_opt $1" func_enable_tag "$1" shift ;; --verbose|-v) opt_quiet=false opt_verbose=: func_append preserve_args " $_G_opt" ;; # An option not handled by this hook function: *) set dummy "$_G_opt" ${1+"$@"}; shift; break ;; esac done # save modified positional parameters for caller func_quote_for_eval ${1+"$@"} libtool_parse_options_result=$func_quote_for_eval_result } func_add_hook func_parse_options libtool_parse_options # libtool_validate_options [ARG]... # --------------------------------- # Perform any sanity checks on option settings and/or unconsumed # arguments. libtool_validate_options () { # save first non-option argument if test 0 -lt $#; then nonopt=$1 shift fi # preserve --debug test : = "$debug_cmd" || func_append preserve_args " --debug" case $host in # Solaris2 added to fix http://debbugs.gnu.org/cgi/bugreport.cgi?bug=16452 # see also: http://gcc.gnu.org/bugzilla/show_bug.cgi?id=59788 *cygwin* | *mingw* | *pw32* | *cegcc* | *solaris2* | *os2*) # don't eliminate duplications in $postdeps and $predeps opt_duplicate_compiler_generated_deps=: ;; *) opt_duplicate_compiler_generated_deps=$opt_preserve_dup_deps ;; esac $opt_help || { # Sanity checks first: func_check_version_match test yes != "$build_libtool_libs" \ && test yes != "$build_old_libs" \ && func_fatal_configuration "not configured to build any kind of library" # Darwin sucks eval std_shrext=\"$shrext_cmds\" # Only execute mode is allowed to have -dlopen flags. if test -n "$opt_dlopen" && test execute != "$opt_mode"; then func_error "unrecognized option '-dlopen'" $ECHO "$help" 1>&2 exit $EXIT_FAILURE fi # Change the help message to a mode-specific one. generic_help=$help help="Try '$progname --help --mode=$opt_mode' for more information." } # Pass back the unparsed argument list func_quote_for_eval ${1+"$@"} libtool_validate_options_result=$func_quote_for_eval_result } func_add_hook func_validate_options libtool_validate_options # Process options as early as possible so that --help and --version # can return quickly. func_options ${1+"$@"} eval set dummy "$func_options_result"; shift ## ----------- ## ## Main. ## ## ----------- ## magic='%%%MAGIC variable%%%' magic_exe='%%%MAGIC EXE variable%%%' # Global variables. extracted_archives= extracted_serial=0 # If this variable is set in any of the actions, the command in it # will be execed at the end. This prevents here-documents from being # left over by shells. exec_cmd= # A function that is used when there is no print builtin or printf. func_fallback_echo () { eval 'cat <<_LTECHO_EOF $1 _LTECHO_EOF' } # func_generated_by_libtool # True iff stdin has been generated by Libtool. This function is only # a basic sanity check; it will hardly flush out determined imposters. func_generated_by_libtool_p () { $GREP "^# Generated by .*$PACKAGE" > /dev/null 2>&1 } # func_lalib_p file # True iff FILE is a libtool '.la' library or '.lo' object file. # This function is only a basic sanity check; it will hardly flush out # determined imposters. func_lalib_p () { test -f "$1" && $SED -e 4q "$1" 2>/dev/null | func_generated_by_libtool_p } # func_lalib_unsafe_p file # True iff FILE is a libtool '.la' library or '.lo' object file. # This function implements the same check as func_lalib_p without # resorting to external programs. To this end, it redirects stdin and # closes it afterwards, without saving the original file descriptor. # As a safety measure, use it only where a negative result would be # fatal anyway. Works if 'file' does not exist. func_lalib_unsafe_p () { lalib_p=no if test -f "$1" && test -r "$1" && exec 5<&0 <"$1"; then for lalib_p_l in 1 2 3 4 do read lalib_p_line case $lalib_p_line in \#\ Generated\ by\ *$PACKAGE* ) lalib_p=yes; break;; esac done exec 0<&5 5<&- fi test yes = "$lalib_p" } # func_ltwrapper_script_p file # True iff FILE is a libtool wrapper script # This function is only a basic sanity check; it will hardly flush out # determined imposters. func_ltwrapper_script_p () { test -f "$1" && $lt_truncate_bin < "$1" 2>/dev/null | func_generated_by_libtool_p } # func_ltwrapper_executable_p file # True iff FILE is a libtool wrapper executable # This function is only a basic sanity check; it will hardly flush out # determined imposters. func_ltwrapper_executable_p () { func_ltwrapper_exec_suffix= case $1 in *.exe) ;; *) func_ltwrapper_exec_suffix=.exe ;; esac $GREP "$magic_exe" "$1$func_ltwrapper_exec_suffix" >/dev/null 2>&1 } # func_ltwrapper_scriptname file # Assumes file is an ltwrapper_executable # uses $file to determine the appropriate filename for a # temporary ltwrapper_script. func_ltwrapper_scriptname () { func_dirname_and_basename "$1" "" "." func_stripname '' '.exe' "$func_basename_result" func_ltwrapper_scriptname_result=$func_dirname_result/$objdir/${func_stripname_result}_ltshwrapper } # func_ltwrapper_p file # True iff FILE is a libtool wrapper script or wrapper executable # This function is only a basic sanity check; it will hardly flush out # determined imposters. func_ltwrapper_p () { func_ltwrapper_script_p "$1" || func_ltwrapper_executable_p "$1" } # func_execute_cmds commands fail_cmd # Execute tilde-delimited COMMANDS. # If FAIL_CMD is given, eval that upon failure. # FAIL_CMD may read-access the current command in variable CMD! func_execute_cmds () { $debug_cmd save_ifs=$IFS; IFS='~' for cmd in $1; do IFS=$sp$nl eval cmd=\"$cmd\" IFS=$save_ifs func_show_eval "$cmd" "${2-:}" done IFS=$save_ifs } # func_source file # Source FILE, adding directory component if necessary. # Note that it is not necessary on cygwin/mingw to append a dot to # FILE even if both FILE and FILE.exe exist: automatic-append-.exe # behavior happens only for exec(3), not for open(2)! Also, sourcing # 'FILE.' does not work on cygwin managed mounts. func_source () { $debug_cmd case $1 in */* | *\\*) . "$1" ;; *) . "./$1" ;; esac } # func_resolve_sysroot PATH # Replace a leading = in PATH with a sysroot. Store the result into # func_resolve_sysroot_result func_resolve_sysroot () { func_resolve_sysroot_result=$1 case $func_resolve_sysroot_result in =*) func_stripname '=' '' "$func_resolve_sysroot_result" func_resolve_sysroot_result=$lt_sysroot$func_stripname_result ;; esac } # func_replace_sysroot PATH # If PATH begins with the sysroot, replace it with = and # store the result into func_replace_sysroot_result. func_replace_sysroot () { case $lt_sysroot:$1 in ?*:"$lt_sysroot"*) func_stripname "$lt_sysroot" '' "$1" func_replace_sysroot_result='='$func_stripname_result ;; *) # Including no sysroot. func_replace_sysroot_result=$1 ;; esac } # func_infer_tag arg # Infer tagged configuration to use if any are available and # if one wasn't chosen via the "--tag" command line option. # Only attempt this if the compiler in the base compile # command doesn't match the default compiler. # arg is usually of the form 'gcc ...' func_infer_tag () { $debug_cmd if test -n "$available_tags" && test -z "$tagname"; then CC_quoted= for arg in $CC; do func_append_quoted CC_quoted "$arg" done CC_expanded=`func_echo_all $CC` CC_quoted_expanded=`func_echo_all $CC_quoted` case $@ in # Blanks in the command may have been stripped by the calling shell, # but not from the CC environment variable when configure was run. " $CC "* | "$CC "* | " $CC_expanded "* | "$CC_expanded "* | \ " $CC_quoted"* | "$CC_quoted "* | " $CC_quoted_expanded "* | "$CC_quoted_expanded "*) ;; # Blanks at the start of $base_compile will cause this to fail # if we don't check for them as well. *) for z in $available_tags; do if $GREP "^# ### BEGIN LIBTOOL TAG CONFIG: $z$" < "$progpath" > /dev/null; then # Evaluate the configuration. eval "`$SED -n -e '/^# ### BEGIN LIBTOOL TAG CONFIG: '$z'$/,/^# ### END LIBTOOL TAG CONFIG: '$z'$/p' < $progpath`" CC_quoted= for arg in $CC; do # Double-quote args containing other shell metacharacters. func_append_quoted CC_quoted "$arg" done CC_expanded=`func_echo_all $CC` CC_quoted_expanded=`func_echo_all $CC_quoted` case "$@ " in " $CC "* | "$CC "* | " $CC_expanded "* | "$CC_expanded "* | \ " $CC_quoted"* | "$CC_quoted "* | " $CC_quoted_expanded "* | "$CC_quoted_expanded "*) # The compiler in the base compile command matches # the one in the tagged configuration. # Assume this is the tagged configuration we want. tagname=$z break ;; esac fi done # If $tagname still isn't set, then no tagged configuration # was found and let the user know that the "--tag" command # line option must be used. if test -z "$tagname"; then func_echo "unable to infer tagged configuration" func_fatal_error "specify a tag with '--tag'" # else # func_verbose "using $tagname tagged configuration" fi ;; esac fi } # func_write_libtool_object output_name pic_name nonpic_name # Create a libtool object file (analogous to a ".la" file), # but don't create it if we're doing a dry run. func_write_libtool_object () { write_libobj=$1 if test yes = "$build_libtool_libs"; then write_lobj=\'$2\' else write_lobj=none fi if test yes = "$build_old_libs"; then write_oldobj=\'$3\' else write_oldobj=none fi $opt_dry_run || { cat >${write_libobj}T <<EOF # $write_libobj - a libtool object file # Generated by $PROGRAM (GNU $PACKAGE) $VERSION # # Please DO NOT delete this file! # It is necessary for linking the library. # Name of the PIC object. pic_object=$write_lobj # Name of the non-PIC object non_pic_object=$write_oldobj EOF $MV "${write_libobj}T" "$write_libobj" } } ################################################## # FILE NAME AND PATH CONVERSION HELPER FUNCTIONS # ################################################## # func_convert_core_file_wine_to_w32 ARG # Helper function used by file name conversion functions when $build is *nix, # and $host is mingw, cygwin, or some other w32 environment. Relies on a # correctly configured wine environment available, with the winepath program # in $build's $PATH. # # ARG is the $build file name to be converted to w32 format. # Result is available in $func_convert_core_file_wine_to_w32_result, and will # be empty on error (or when ARG is empty) func_convert_core_file_wine_to_w32 () { $debug_cmd func_convert_core_file_wine_to_w32_result=$1 if test -n "$1"; then # Unfortunately, winepath does not exit with a non-zero error code, so we # are forced to check the contents of stdout. On the other hand, if the # command is not found, the shell will set an exit code of 127 and print # *an error message* to stdout. So we must check for both error code of # zero AND non-empty stdout, which explains the odd construction: func_convert_core_file_wine_to_w32_tmp=`winepath -w "$1" 2>/dev/null` if test "$?" -eq 0 && test -n "$func_convert_core_file_wine_to_w32_tmp"; then func_convert_core_file_wine_to_w32_result=`$ECHO "$func_convert_core_file_wine_to_w32_tmp" | $SED -e "$sed_naive_backslashify"` else func_convert_core_file_wine_to_w32_result= fi fi } # end: func_convert_core_file_wine_to_w32 # func_convert_core_path_wine_to_w32 ARG # Helper function used by path conversion functions when $build is *nix, and # $host is mingw, cygwin, or some other w32 environment. Relies on a correctly # configured wine environment available, with the winepath program in $build's # $PATH. Assumes ARG has no leading or trailing path separator characters. # # ARG is path to be converted from $build format to win32. # Result is available in $func_convert_core_path_wine_to_w32_result. # Unconvertible file (directory) names in ARG are skipped; if no directory names # are convertible, then the result may be empty. func_convert_core_path_wine_to_w32 () { $debug_cmd # unfortunately, winepath doesn't convert paths, only file names func_convert_core_path_wine_to_w32_result= if test -n "$1"; then oldIFS=$IFS IFS=: for func_convert_core_path_wine_to_w32_f in $1; do IFS=$oldIFS func_convert_core_file_wine_to_w32 "$func_convert_core_path_wine_to_w32_f" if test -n "$func_convert_core_file_wine_to_w32_result"; then if test -z "$func_convert_core_path_wine_to_w32_result"; then func_convert_core_path_wine_to_w32_result=$func_convert_core_file_wine_to_w32_result else func_append func_convert_core_path_wine_to_w32_result ";$func_convert_core_file_wine_to_w32_result" fi fi done IFS=$oldIFS fi } # end: func_convert_core_path_wine_to_w32 # func_cygpath ARGS... # Wrapper around calling the cygpath program via LT_CYGPATH. This is used when # when (1) $build is *nix and Cygwin is hosted via a wine environment; or (2) # $build is MSYS and $host is Cygwin, or (3) $build is Cygwin. In case (1) or # (2), returns the Cygwin file name or path in func_cygpath_result (input # file name or path is assumed to be in w32 format, as previously converted # from $build's *nix or MSYS format). In case (3), returns the w32 file name # or path in func_cygpath_result (input file name or path is assumed to be in # Cygwin format). Returns an empty string on error. # # ARGS are passed to cygpath, with the last one being the file name or path to # be converted. # # Specify the absolute *nix (or w32) name to cygpath in the LT_CYGPATH # environment variable; do not put it in $PATH. func_cygpath () { $debug_cmd if test -n "$LT_CYGPATH" && test -f "$LT_CYGPATH"; then func_cygpath_result=`$LT_CYGPATH "$@" 2>/dev/null` if test "$?" -ne 0; then # on failure, ensure result is empty func_cygpath_result= fi else func_cygpath_result= func_error "LT_CYGPATH is empty or specifies non-existent file: '$LT_CYGPATH'" fi } #end: func_cygpath # func_convert_core_msys_to_w32 ARG # Convert file name or path ARG from MSYS format to w32 format. Return # result in func_convert_core_msys_to_w32_result. func_convert_core_msys_to_w32 () { $debug_cmd # awkward: cmd appends spaces to result func_convert_core_msys_to_w32_result=`( cmd //c echo "$1" ) 2>/dev/null | $SED -e 's/[ ]*$//' -e "$sed_naive_backslashify"` } #end: func_convert_core_msys_to_w32 # func_convert_file_check ARG1 ARG2 # Verify that ARG1 (a file name in $build format) was converted to $host # format in ARG2. Otherwise, emit an error message, but continue (resetting # func_to_host_file_result to ARG1). func_convert_file_check () { $debug_cmd if test -z "$2" && test -n "$1"; then func_error "Could not determine host file name corresponding to" func_error " '$1'" func_error "Continuing, but uninstalled executables may not work." # Fallback: func_to_host_file_result=$1 fi } # end func_convert_file_check # func_convert_path_check FROM_PATHSEP TO_PATHSEP FROM_PATH TO_PATH # Verify that FROM_PATH (a path in $build format) was converted to $host # format in TO_PATH. Otherwise, emit an error message, but continue, resetting # func_to_host_file_result to a simplistic fallback value (see below). func_convert_path_check () { $debug_cmd if test -z "$4" && test -n "$3"; then func_error "Could not determine the host path corresponding to" func_error " '$3'" func_error "Continuing, but uninstalled executables may not work." # Fallback. This is a deliberately simplistic "conversion" and # should not be "improved". See libtool.info. if test "x$1" != "x$2"; then lt_replace_pathsep_chars="s|$1|$2|g" func_to_host_path_result=`echo "$3" | $SED -e "$lt_replace_pathsep_chars"` else func_to_host_path_result=$3 fi fi } # end func_convert_path_check # func_convert_path_front_back_pathsep FRONTPAT BACKPAT REPL ORIG # Modifies func_to_host_path_result by prepending REPL if ORIG matches FRONTPAT # and appending REPL if ORIG matches BACKPAT. func_convert_path_front_back_pathsep () { $debug_cmd case $4 in $1 ) func_to_host_path_result=$3$func_to_host_path_result ;; esac case $4 in $2 ) func_append func_to_host_path_result "$3" ;; esac } # end func_convert_path_front_back_pathsep ################################################## # $build to $host FILE NAME CONVERSION FUNCTIONS # ################################################## # invoked via '$to_host_file_cmd ARG' # # In each case, ARG is the path to be converted from $build to $host format. # Result will be available in $func_to_host_file_result. # func_to_host_file ARG # Converts the file name ARG from $build format to $host format. Return result # in func_to_host_file_result. func_to_host_file () { $debug_cmd $to_host_file_cmd "$1" } # end func_to_host_file # func_to_tool_file ARG LAZY # converts the file name ARG from $build format to toolchain format. Return # result in func_to_tool_file_result. If the conversion in use is listed # in (the comma separated) LAZY, no conversion takes place. func_to_tool_file () { $debug_cmd case ,$2, in *,"$to_tool_file_cmd",*) func_to_tool_file_result=$1 ;; *) $to_tool_file_cmd "$1" func_to_tool_file_result=$func_to_host_file_result ;; esac } # end func_to_tool_file # func_convert_file_noop ARG # Copy ARG to func_to_host_file_result. func_convert_file_noop () { func_to_host_file_result=$1 } # end func_convert_file_noop # func_convert_file_msys_to_w32 ARG # Convert file name ARG from (mingw) MSYS to (mingw) w32 format; automatic # conversion to w32 is not available inside the cwrapper. Returns result in # func_to_host_file_result. func_convert_file_msys_to_w32 () { $debug_cmd func_to_host_file_result=$1 if test -n "$1"; then func_convert_core_msys_to_w32 "$1" func_to_host_file_result=$func_convert_core_msys_to_w32_result fi func_convert_file_check "$1" "$func_to_host_file_result" } # end func_convert_file_msys_to_w32 # func_convert_file_cygwin_to_w32 ARG # Convert file name ARG from Cygwin to w32 format. Returns result in # func_to_host_file_result. func_convert_file_cygwin_to_w32 () { $debug_cmd func_to_host_file_result=$1 if test -n "$1"; then # because $build is cygwin, we call "the" cygpath in $PATH; no need to use # LT_CYGPATH in this case. func_to_host_file_result=`cygpath -m "$1"` fi func_convert_file_check "$1" "$func_to_host_file_result" } # end func_convert_file_cygwin_to_w32 # func_convert_file_nix_to_w32 ARG # Convert file name ARG from *nix to w32 format. Requires a wine environment # and a working winepath. Returns result in func_to_host_file_result. func_convert_file_nix_to_w32 () { $debug_cmd func_to_host_file_result=$1 if test -n "$1"; then func_convert_core_file_wine_to_w32 "$1" func_to_host_file_result=$func_convert_core_file_wine_to_w32_result fi func_convert_file_check "$1" "$func_to_host_file_result" } # end func_convert_file_nix_to_w32 # func_convert_file_msys_to_cygwin ARG # Convert file name ARG from MSYS to Cygwin format. Requires LT_CYGPATH set. # Returns result in func_to_host_file_result. func_convert_file_msys_to_cygwin () { $debug_cmd func_to_host_file_result=$1 if test -n "$1"; then func_convert_core_msys_to_w32 "$1" func_cygpath -u "$func_convert_core_msys_to_w32_result" func_to_host_file_result=$func_cygpath_result fi func_convert_file_check "$1" "$func_to_host_file_result" } # end func_convert_file_msys_to_cygwin # func_convert_file_nix_to_cygwin ARG # Convert file name ARG from *nix to Cygwin format. Requires Cygwin installed # in a wine environment, working winepath, and LT_CYGPATH set. Returns result # in func_to_host_file_result. func_convert_file_nix_to_cygwin () { $debug_cmd func_to_host_file_result=$1 if test -n "$1"; then # convert from *nix to w32, then use cygpath to convert from w32 to cygwin. func_convert_core_file_wine_to_w32 "$1" func_cygpath -u "$func_convert_core_file_wine_to_w32_result" func_to_host_file_result=$func_cygpath_result fi func_convert_file_check "$1" "$func_to_host_file_result" } # end func_convert_file_nix_to_cygwin ############################################# # $build to $host PATH CONVERSION FUNCTIONS # ############################################# # invoked via '$to_host_path_cmd ARG' # # In each case, ARG is the path to be converted from $build to $host format. # The result will be available in $func_to_host_path_result. # # Path separators are also converted from $build format to $host format. If # ARG begins or ends with a path separator character, it is preserved (but # converted to $host format) on output. # # All path conversion functions are named using the following convention: # file name conversion function : func_convert_file_X_to_Y () # path conversion function : func_convert_path_X_to_Y () # where, for any given $build/$host combination the 'X_to_Y' value is the # same. If conversion functions are added for new $build/$host combinations, # the two new functions must follow this pattern, or func_init_to_host_path_cmd # will break. # func_init_to_host_path_cmd # Ensures that function "pointer" variable $to_host_path_cmd is set to the # appropriate value, based on the value of $to_host_file_cmd. to_host_path_cmd= func_init_to_host_path_cmd () { $debug_cmd if test -z "$to_host_path_cmd"; then func_stripname 'func_convert_file_' '' "$to_host_file_cmd" to_host_path_cmd=func_convert_path_$func_stripname_result fi } # func_to_host_path ARG # Converts the path ARG from $build format to $host format. Return result # in func_to_host_path_result. func_to_host_path () { $debug_cmd func_init_to_host_path_cmd $to_host_path_cmd "$1" } # end func_to_host_path # func_convert_path_noop ARG # Copy ARG to func_to_host_path_result. func_convert_path_noop () { func_to_host_path_result=$1 } # end func_convert_path_noop # func_convert_path_msys_to_w32 ARG # Convert path ARG from (mingw) MSYS to (mingw) w32 format; automatic # conversion to w32 is not available inside the cwrapper. Returns result in # func_to_host_path_result. func_convert_path_msys_to_w32 () { $debug_cmd func_to_host_path_result=$1 if test -n "$1"; then # Remove leading and trailing path separator characters from ARG. MSYS # behavior is inconsistent here; cygpath turns them into '.;' and ';.'; # and winepath ignores them completely. func_stripname : : "$1" func_to_host_path_tmp1=$func_stripname_result func_convert_core_msys_to_w32 "$func_to_host_path_tmp1" func_to_host_path_result=$func_convert_core_msys_to_w32_result func_convert_path_check : ";" \ "$func_to_host_path_tmp1" "$func_to_host_path_result" func_convert_path_front_back_pathsep ":*" "*:" ";" "$1" fi } # end func_convert_path_msys_to_w32 # func_convert_path_cygwin_to_w32 ARG # Convert path ARG from Cygwin to w32 format. Returns result in # func_to_host_file_result. func_convert_path_cygwin_to_w32 () { $debug_cmd func_to_host_path_result=$1 if test -n "$1"; then # See func_convert_path_msys_to_w32: func_stripname : : "$1" func_to_host_path_tmp1=$func_stripname_result func_to_host_path_result=`cygpath -m -p "$func_to_host_path_tmp1"` func_convert_path_check : ";" \ "$func_to_host_path_tmp1" "$func_to_host_path_result" func_convert_path_front_back_pathsep ":*" "*:" ";" "$1" fi } # end func_convert_path_cygwin_to_w32 # func_convert_path_nix_to_w32 ARG # Convert path ARG from *nix to w32 format. Requires a wine environment and # a working winepath. Returns result in func_to_host_file_result. func_convert_path_nix_to_w32 () { $debug_cmd func_to_host_path_result=$1 if test -n "$1"; then # See func_convert_path_msys_to_w32: func_stripname : : "$1" func_to_host_path_tmp1=$func_stripname_result func_convert_core_path_wine_to_w32 "$func_to_host_path_tmp1" func_to_host_path_result=$func_convert_core_path_wine_to_w32_result func_convert_path_check : ";" \ "$func_to_host_path_tmp1" "$func_to_host_path_result" func_convert_path_front_back_pathsep ":*" "*:" ";" "$1" fi } # end func_convert_path_nix_to_w32 # func_convert_path_msys_to_cygwin ARG # Convert path ARG from MSYS to Cygwin format. Requires LT_CYGPATH set. # Returns result in func_to_host_file_result. func_convert_path_msys_to_cygwin () { $debug_cmd func_to_host_path_result=$1 if test -n "$1"; then # See func_convert_path_msys_to_w32: func_stripname : : "$1" func_to_host_path_tmp1=$func_stripname_result func_convert_core_msys_to_w32 "$func_to_host_path_tmp1" func_cygpath -u -p "$func_convert_core_msys_to_w32_result" func_to_host_path_result=$func_cygpath_result func_convert_path_check : : \ "$func_to_host_path_tmp1" "$func_to_host_path_result" func_convert_path_front_back_pathsep ":*" "*:" : "$1" fi } # end func_convert_path_msys_to_cygwin # func_convert_path_nix_to_cygwin ARG # Convert path ARG from *nix to Cygwin format. Requires Cygwin installed in a # a wine environment, working winepath, and LT_CYGPATH set. Returns result in # func_to_host_file_result. func_convert_path_nix_to_cygwin () { $debug_cmd func_to_host_path_result=$1 if test -n "$1"; then # Remove leading and trailing path separator characters from # ARG. msys behavior is inconsistent here, cygpath turns them # into '.;' and ';.', and winepath ignores them completely. func_stripname : : "$1" func_to_host_path_tmp1=$func_stripname_result func_convert_core_path_wine_to_w32 "$func_to_host_path_tmp1" func_cygpath -u -p "$func_convert_core_path_wine_to_w32_result" func_to_host_path_result=$func_cygpath_result func_convert_path_check : : \ "$func_to_host_path_tmp1" "$func_to_host_path_result" func_convert_path_front_back_pathsep ":*" "*:" : "$1" fi } # end func_convert_path_nix_to_cygwin # func_dll_def_p FILE # True iff FILE is a Windows DLL '.def' file. # Keep in sync with _LT_DLL_DEF_P in libtool.m4 func_dll_def_p () { $debug_cmd func_dll_def_p_tmp=`$SED -n \ -e 's/^[ ]*//' \ -e '/^\(;.*\)*$/d' \ -e 's/^\(EXPORTS\|LIBRARY\)\([ ].*\)*$/DEF/p' \ -e q \ "$1"` test DEF = "$func_dll_def_p_tmp" } # func_mode_compile arg... func_mode_compile () { $debug_cmd # Get the compilation command and the source file. base_compile= srcfile=$nonopt # always keep a non-empty value in "srcfile" suppress_opt=yes suppress_output= arg_mode=normal libobj= later= pie_flag= for arg do case $arg_mode in arg ) # do not "continue". Instead, add this to base_compile lastarg=$arg arg_mode=normal ;; target ) libobj=$arg arg_mode=normal continue ;; normal ) # Accept any command-line options. case $arg in -o) test -n "$libobj" && \ func_fatal_error "you cannot specify '-o' more than once" arg_mode=target continue ;; -pie | -fpie | -fPIE) func_append pie_flag " $arg" continue ;; -shared | -static | -prefer-pic | -prefer-non-pic) func_append later " $arg" continue ;; -no-suppress) suppress_opt=no continue ;; -Xcompiler) arg_mode=arg # the next one goes into the "base_compile" arg list continue # The current "srcfile" will either be retained or ;; # replaced later. I would guess that would be a bug. -Wc,*) func_stripname '-Wc,' '' "$arg" args=$func_stripname_result lastarg= save_ifs=$IFS; IFS=, for arg in $args; do IFS=$save_ifs func_append_quoted lastarg "$arg" done IFS=$save_ifs func_stripname ' ' '' "$lastarg" lastarg=$func_stripname_result # Add the arguments to base_compile. func_append base_compile " $lastarg" continue ;; *) # Accept the current argument as the source file. # The previous "srcfile" becomes the current argument. # lastarg=$srcfile srcfile=$arg ;; esac # case $arg ;; esac # case $arg_mode # Aesthetically quote the previous argument. func_append_quoted base_compile "$lastarg" done # for arg case $arg_mode in arg) func_fatal_error "you must specify an argument for -Xcompile" ;; target) func_fatal_error "you must specify a target with '-o'" ;; *) # Get the name of the library object. test -z "$libobj" && { func_basename "$srcfile" libobj=$func_basename_result } ;; esac # Recognize several different file suffixes. # If the user specifies -o file.o, it is replaced with file.lo case $libobj in *.[cCFSifmso] | \ *.ada | *.adb | *.ads | *.asm | \ *.c++ | *.cc | *.ii | *.class | *.cpp | *.cxx | \ *.[fF][09]? | *.for | *.java | *.go | *.obj | *.sx | *.cu | *.cup) func_xform "$libobj" libobj=$func_xform_result ;; esac case $libobj in *.lo) func_lo2o "$libobj"; obj=$func_lo2o_result ;; *) func_fatal_error "cannot determine name of library object from '$libobj'" ;; esac func_infer_tag $base_compile for arg in $later; do case $arg in -shared) test yes = "$build_libtool_libs" \ || func_fatal_configuration "cannot build a shared library" build_old_libs=no continue ;; -static) build_libtool_libs=no build_old_libs=yes continue ;; -prefer-pic) pic_mode=yes continue ;; -prefer-non-pic) pic_mode=no continue ;; esac done func_quote_for_eval "$libobj" test "X$libobj" != "X$func_quote_for_eval_result" \ && $ECHO "X$libobj" | $GREP '[]~#^*{};<>?"'"'"' &()|`$[]' \ && func_warning "libobj name '$libobj' may not contain shell special characters." func_dirname_and_basename "$obj" "/" "" objname=$func_basename_result xdir=$func_dirname_result lobj=$xdir$objdir/$objname test -z "$base_compile" && \ func_fatal_help "you must specify a compilation command" # Delete any leftover library objects. if test yes = "$build_old_libs"; then removelist="$obj $lobj $libobj ${libobj}T" else removelist="$lobj $libobj ${libobj}T" fi # On Cygwin there's no "real" PIC flag so we must build both object types case $host_os in cygwin* | mingw* | pw32* | os2* | cegcc*) pic_mode=default ;; esac if test no = "$pic_mode" && test pass_all != "$deplibs_check_method"; then # non-PIC code in shared libraries is not supported pic_mode=default fi # Calculate the filename of the output object if compiler does # not support -o with -c if test no = "$compiler_c_o"; then output_obj=`$ECHO "$srcfile" | $SED 's%^.*/%%; s%\.[^.]*$%%'`.$objext lockfile=$output_obj.lock else output_obj= need_locks=no lockfile= fi # Lock this critical section if it is needed # We use this script file to make the link, it avoids creating a new file if test yes = "$need_locks"; then until $opt_dry_run || ln "$progpath" "$lockfile" 2>/dev/null; do func_echo "Waiting for $lockfile to be removed" sleep 2 done elif test warn = "$need_locks"; then if test -f "$lockfile"; then $ECHO "\ *** ERROR, $lockfile exists and contains: `cat $lockfile 2>/dev/null` This indicates that another process is trying to use the same temporary object file, and libtool could not work around it because your compiler does not support '-c' and '-o' together. If you repeat this compilation, it may succeed, by chance, but you had better avoid parallel builds (make -j) in this platform, or get a better compiler." $opt_dry_run || $RM $removelist exit $EXIT_FAILURE fi func_append removelist " $output_obj" $ECHO "$srcfile" > "$lockfile" fi $opt_dry_run || $RM $removelist func_append removelist " $lockfile" trap '$opt_dry_run || $RM $removelist; exit $EXIT_FAILURE' 1 2 15 func_to_tool_file "$srcfile" func_convert_file_msys_to_w32 srcfile=$func_to_tool_file_result func_quote_for_eval "$srcfile" qsrcfile=$func_quote_for_eval_result # Only build a PIC object if we are building libtool libraries. if test yes = "$build_libtool_libs"; then # Without this assignment, base_compile gets emptied. fbsd_hideous_sh_bug=$base_compile if test no != "$pic_mode"; then command="$base_compile $qsrcfile $pic_flag" else # Don't build PIC code command="$base_compile $qsrcfile" fi func_mkdir_p "$xdir$objdir" if test -z "$output_obj"; then # Place PIC objects in $objdir func_append command " -o $lobj" fi func_show_eval_locale "$command" \ 'test -n "$output_obj" && $RM $removelist; exit $EXIT_FAILURE' if test warn = "$need_locks" && test "X`cat $lockfile 2>/dev/null`" != "X$srcfile"; then $ECHO "\ *** ERROR, $lockfile contains: `cat $lockfile 2>/dev/null` but it should contain: $srcfile This indicates that another process is trying to use the same temporary object file, and libtool could not work around it because your compiler does not support '-c' and '-o' together. If you repeat this compilation, it may succeed, by chance, but you had better avoid parallel builds (make -j) in this platform, or get a better compiler." $opt_dry_run || $RM $removelist exit $EXIT_FAILURE fi # Just move the object if needed, then go on to compile the next one if test -n "$output_obj" && test "X$output_obj" != "X$lobj"; then func_show_eval '$MV "$output_obj" "$lobj"' \ 'error=$?; $opt_dry_run || $RM $removelist; exit $error' fi # Allow error messages only from the first compilation. if test yes = "$suppress_opt"; then suppress_output=' >/dev/null 2>&1' fi fi # Only build a position-dependent object if we build old libraries. if test yes = "$build_old_libs"; then if test yes != "$pic_mode"; then # Don't build PIC code command="$base_compile $qsrcfile$pie_flag" else command="$base_compile $qsrcfile $pic_flag" fi if test yes = "$compiler_c_o"; then func_append command " -o $obj" fi # Suppress compiler output if we already did a PIC compilation. func_append command "$suppress_output" func_show_eval_locale "$command" \ '$opt_dry_run || $RM $removelist; exit $EXIT_FAILURE' if test warn = "$need_locks" && test "X`cat $lockfile 2>/dev/null`" != "X$srcfile"; then $ECHO "\ *** ERROR, $lockfile contains: `cat $lockfile 2>/dev/null` but it should contain: $srcfile This indicates that another process is trying to use the same temporary object file, and libtool could not work around it because your compiler does not support '-c' and '-o' together. If you repeat this compilation, it may succeed, by chance, but you had better avoid parallel builds (make -j) in this platform, or get a better compiler." $opt_dry_run || $RM $removelist exit $EXIT_FAILURE fi # Just move the object if needed if test -n "$output_obj" && test "X$output_obj" != "X$obj"; then func_show_eval '$MV "$output_obj" "$obj"' \ 'error=$?; $opt_dry_run || $RM $removelist; exit $error' fi fi $opt_dry_run || { func_write_libtool_object "$libobj" "$objdir/$objname" "$objname" # Unlock the critical section if it was locked if test no != "$need_locks"; then removelist=$lockfile $RM "$lockfile" fi } exit $EXIT_SUCCESS } $opt_help || { test compile = "$opt_mode" && func_mode_compile ${1+"$@"} } func_mode_help () { # We need to display help for each of the modes. case $opt_mode in "") # Generic help is extracted from the usage comments # at the start of this file. func_help ;; clean) $ECHO \ "Usage: $progname [OPTION]... --mode=clean RM [RM-OPTION]... FILE... Remove files from the build directory. RM is the name of the program to use to delete files associated with each FILE (typically '/bin/rm'). RM-OPTIONS are options (such as '-f') to be passed to RM. If FILE is a libtool library, object or program, all the files associated with it are deleted. Otherwise, only FILE itself is deleted using RM." ;; compile) $ECHO \ "Usage: $progname [OPTION]... --mode=compile COMPILE-COMMAND... SOURCEFILE Compile a source file into a libtool library object. This mode accepts the following additional options: -o OUTPUT-FILE set the output file name to OUTPUT-FILE -no-suppress do not suppress compiler output for multiple passes -prefer-pic try to build PIC objects only -prefer-non-pic try to build non-PIC objects only -shared do not build a '.o' file suitable for static linking -static only build a '.o' file suitable for static linking -Wc,FLAG pass FLAG directly to the compiler COMPILE-COMMAND is a command to be used in creating a 'standard' object file from the given SOURCEFILE. The output file name is determined by removing the directory component from SOURCEFILE, then substituting the C source code suffix '.c' with the library object suffix, '.lo'." ;; execute) $ECHO \ "Usage: $progname [OPTION]... --mode=execute COMMAND [ARGS]... Automatically set library path, then run a program. This mode accepts the following additional options: -dlopen FILE add the directory containing FILE to the library path This mode sets the library path environment variable according to '-dlopen' flags. If any of the ARGS are libtool executable wrappers, then they are translated into their corresponding uninstalled binary, and any of their required library directories are added to the library path. Then, COMMAND is executed, with ARGS as arguments." ;; finish) $ECHO \ "Usage: $progname [OPTION]... --mode=finish [LIBDIR]... Complete the installation of libtool libraries. Each LIBDIR is a directory that contains libtool libraries. The commands that this mode executes may require superuser privileges. Use the '--dry-run' option if you just want to see what would be executed." ;; install) $ECHO \ "Usage: $progname [OPTION]... --mode=install INSTALL-COMMAND... Install executables or libraries. INSTALL-COMMAND is the installation command. The first component should be either the 'install' or 'cp' program. The following components of INSTALL-COMMAND are treated specially: -inst-prefix-dir PREFIX-DIR Use PREFIX-DIR as a staging area for installation The rest of the components are interpreted as arguments to that command (only BSD-compatible install options are recognized)." ;; link) $ECHO \ "Usage: $progname [OPTION]... --mode=link LINK-COMMAND... Link object files or libraries together to form another library, or to create an executable program. LINK-COMMAND is a command using the C compiler that you would use to create a program from several object files. The following components of LINK-COMMAND are treated specially: -all-static do not do any dynamic linking at all -avoid-version do not add a version suffix if possible -bindir BINDIR specify path to binaries directory (for systems where libraries must be found in the PATH setting at runtime) -dlopen FILE '-dlpreopen' FILE if it cannot be dlopened at runtime -dlpreopen FILE link in FILE and add its symbols to lt_preloaded_symbols -export-dynamic allow symbols from OUTPUT-FILE to be resolved with dlsym(3) -export-symbols SYMFILE try to export only the symbols listed in SYMFILE -export-symbols-regex REGEX try to export only the symbols matching REGEX -LLIBDIR search LIBDIR for required installed libraries -lNAME OUTPUT-FILE requires the installed library libNAME -module build a library that can dlopened -no-fast-install disable the fast-install mode -no-install link a not-installable executable -no-undefined declare that a library does not refer to external symbols -o OUTPUT-FILE create OUTPUT-FILE from the specified objects -objectlist FILE use a list of object files found in FILE to specify objects -os2dllname NAME force a short DLL name on OS/2 (no effect on other OSes) -precious-files-regex REGEX don't remove output files matching REGEX -release RELEASE specify package release information -rpath LIBDIR the created library will eventually be installed in LIBDIR -R[ ]LIBDIR add LIBDIR to the runtime path of programs and libraries -shared only do dynamic linking of libtool libraries -shrext SUFFIX override the standard shared library file extension -static do not do any dynamic linking of uninstalled libtool libraries -static-libtool-libs do not do any dynamic linking of libtool libraries -version-info CURRENT[:REVISION[:AGE]] specify library version info [each variable defaults to 0] -weak LIBNAME declare that the target provides the LIBNAME interface -Wc,FLAG -Xcompiler FLAG pass linker-specific FLAG directly to the compiler -Wl,FLAG -Xlinker FLAG pass linker-specific FLAG directly to the linker -XCClinker FLAG pass link-specific FLAG to the compiler driver (CC) All other options (arguments beginning with '-') are ignored. Every other argument is treated as a filename. Files ending in '.la' are treated as uninstalled libtool libraries, other files are standard or library object files. If the OUTPUT-FILE ends in '.la', then a libtool library is created, only library objects ('.lo' files) may be specified, and '-rpath' is required, except when creating a convenience library. If OUTPUT-FILE ends in '.a' or '.lib', then a standard library is created using 'ar' and 'ranlib', or on Windows using 'lib'. If OUTPUT-FILE ends in '.lo' or '.$objext', then a reloadable object file is created, otherwise an executable program is created." ;; uninstall) $ECHO \ "Usage: $progname [OPTION]... --mode=uninstall RM [RM-OPTION]... FILE... Remove libraries from an installation directory. RM is the name of the program to use to delete files associated with each FILE (typically '/bin/rm'). RM-OPTIONS are options (such as '-f') to be passed to RM. If FILE is a libtool library, all the files associated with it are deleted. Otherwise, only FILE itself is deleted using RM." ;; *) func_fatal_help "invalid operation mode '$opt_mode'" ;; esac echo $ECHO "Try '$progname --help' for more information about other modes." } # Now that we've collected a possible --mode arg, show help if necessary if $opt_help; then if test : = "$opt_help"; then func_mode_help else { func_help noexit for opt_mode in compile link execute install finish uninstall clean; do func_mode_help done } | $SED -n '1p; 2,$s/^Usage:/ or: /p' { func_help noexit for opt_mode in compile link execute install finish uninstall clean; do echo func_mode_help done } | $SED '1d /^When reporting/,/^Report/{ H d } $x /information about other modes/d /more detailed .*MODE/d s/^Usage:.*--mode=\([^ ]*\) .*/Description of \1 mode:/' fi exit $? fi # func_mode_execute arg... func_mode_execute () { $debug_cmd # The first argument is the command name. cmd=$nonopt test -z "$cmd" && \ func_fatal_help "you must specify a COMMAND" # Handle -dlopen flags immediately. for file in $opt_dlopen; do test -f "$file" \ || func_fatal_help "'$file' is not a file" dir= case $file in *.la) func_resolve_sysroot "$file" file=$func_resolve_sysroot_result # Check to see that this really is a libtool archive. func_lalib_unsafe_p "$file" \ || func_fatal_help "'$lib' is not a valid libtool archive" # Read the libtool library. dlname= library_names= func_source "$file" # Skip this library if it cannot be dlopened. if test -z "$dlname"; then # Warn if it was a shared library. test -n "$library_names" && \ func_warning "'$file' was not linked with '-export-dynamic'" continue fi func_dirname "$file" "" "." dir=$func_dirname_result if test -f "$dir/$objdir/$dlname"; then func_append dir "/$objdir" else if test ! -f "$dir/$dlname"; then func_fatal_error "cannot find '$dlname' in '$dir' or '$dir/$objdir'" fi fi ;; *.lo) # Just add the directory containing the .lo file. func_dirname "$file" "" "." dir=$func_dirname_result ;; *) func_warning "'-dlopen' is ignored for non-libtool libraries and objects" continue ;; esac # Get the absolute pathname. absdir=`cd "$dir" && pwd` test -n "$absdir" && dir=$absdir # Now add the directory to shlibpath_var. if eval "test -z \"\$$shlibpath_var\""; then eval "$shlibpath_var=\"\$dir\"" else eval "$shlibpath_var=\"\$dir:\$$shlibpath_var\"" fi done # This variable tells wrapper scripts just to set shlibpath_var # rather than running their programs. libtool_execute_magic=$magic # Check if any of the arguments is a wrapper script. args= for file do case $file in -* | *.la | *.lo ) ;; *) # Do a test to see if this is really a libtool program. if func_ltwrapper_script_p "$file"; then func_source "$file" # Transform arg to wrapped name. file=$progdir/$program elif func_ltwrapper_executable_p "$file"; then func_ltwrapper_scriptname "$file" func_source "$func_ltwrapper_scriptname_result" # Transform arg to wrapped name. file=$progdir/$program fi ;; esac # Quote arguments (to preserve shell metacharacters). func_append_quoted args "$file" done if $opt_dry_run; then # Display what would be done. if test -n "$shlibpath_var"; then eval "\$ECHO \"\$shlibpath_var=\$$shlibpath_var\"" echo "export $shlibpath_var" fi $ECHO "$cmd$args" exit $EXIT_SUCCESS else if test -n "$shlibpath_var"; then # Export the shlibpath_var. eval "export $shlibpath_var" fi # Restore saved environment variables for lt_var in LANG LANGUAGE LC_ALL LC_CTYPE LC_COLLATE LC_MESSAGES do eval "if test \"\${save_$lt_var+set}\" = set; then $lt_var=\$save_$lt_var; export $lt_var else $lt_unset $lt_var fi" done # Now prepare to actually exec the command. exec_cmd=\$cmd$args fi } test execute = "$opt_mode" && func_mode_execute ${1+"$@"} # func_mode_finish arg... func_mode_finish () { $debug_cmd libs= libdirs= admincmds= for opt in "$nonopt" ${1+"$@"} do if test -d "$opt"; then func_append libdirs " $opt" elif test -f "$opt"; then if func_lalib_unsafe_p "$opt"; then func_append libs " $opt" else func_warning "'$opt' is not a valid libtool archive" fi else func_fatal_error "invalid argument '$opt'" fi done if test -n "$libs"; then if test -n "$lt_sysroot"; then sysroot_regex=`$ECHO "$lt_sysroot" | $SED "$sed_make_literal_regex"` sysroot_cmd="s/\([ ']\)$sysroot_regex/\1/g;" else sysroot_cmd= fi # Remove sysroot references if $opt_dry_run; then for lib in $libs; do echo "removing references to $lt_sysroot and '=' prefixes from $lib" done else tmpdir=`func_mktempdir` for lib in $libs; do $SED -e "$sysroot_cmd s/\([ ']-[LR]\)=/\1/g; s/\([ ']\)=/\1/g" $lib \ > $tmpdir/tmp-la mv -f $tmpdir/tmp-la $lib done ${RM}r "$tmpdir" fi fi if test -n "$finish_cmds$finish_eval" && test -n "$libdirs"; then for libdir in $libdirs; do if test -n "$finish_cmds"; then # Do each command in the finish commands. func_execute_cmds "$finish_cmds" 'admincmds="$admincmds '"$cmd"'"' fi if test -n "$finish_eval"; then # Do the single finish_eval. eval cmds=\"$finish_eval\" $opt_dry_run || eval "$cmds" || func_append admincmds " $cmds" fi done fi # Exit here if they wanted silent mode. $opt_quiet && exit $EXIT_SUCCESS if test -n "$finish_cmds$finish_eval" && test -n "$libdirs"; then echo "----------------------------------------------------------------------" echo "Libraries have been installed in:" for libdir in $libdirs; do $ECHO " $libdir" done echo echo "If you ever happen to want to link against installed libraries" echo "in a given directory, LIBDIR, you must either use libtool, and" echo "specify the full pathname of the library, or use the '-LLIBDIR'" echo "flag during linking and do at least one of the following:" if test -n "$shlibpath_var"; then echo " - add LIBDIR to the '$shlibpath_var' environment variable" echo " during execution" fi if test -n "$runpath_var"; then echo " - add LIBDIR to the '$runpath_var' environment variable" echo " during linking" fi if test -n "$hardcode_libdir_flag_spec"; then libdir=LIBDIR eval flag=\"$hardcode_libdir_flag_spec\" $ECHO " - use the '$flag' linker flag" fi if test -n "$admincmds"; then $ECHO " - have your system administrator run these commands:$admincmds" fi if test -f /etc/ld.so.conf; then echo " - have your system administrator add LIBDIR to '/etc/ld.so.conf'" fi echo echo "See any operating system documentation about shared libraries for" case $host in solaris2.[6789]|solaris2.1[0-9]) echo "more information, such as the ld(1), crle(1) and ld.so(8) manual" echo "pages." ;; *) echo "more information, such as the ld(1) and ld.so(8) manual pages." ;; esac echo "----------------------------------------------------------------------" fi exit $EXIT_SUCCESS } test finish = "$opt_mode" && func_mode_finish ${1+"$@"} # func_mode_install arg... func_mode_install () { $debug_cmd # There may be an optional sh(1) argument at the beginning of # install_prog (especially on Windows NT). if test "$SHELL" = "$nonopt" || test /bin/sh = "$nonopt" || # Allow the use of GNU shtool's install command. case $nonopt in *shtool*) :;; *) false;; esac then # Aesthetically quote it. func_quote_for_eval "$nonopt" install_prog="$func_quote_for_eval_result " arg=$1 shift else install_prog= arg=$nonopt fi # The real first argument should be the name of the installation program. # Aesthetically quote it. func_quote_for_eval "$arg" func_append install_prog "$func_quote_for_eval_result" install_shared_prog=$install_prog case " $install_prog " in *[\\\ /]cp\ *) install_cp=: ;; *) install_cp=false ;; esac # We need to accept at least all the BSD install flags. dest= files= opts= prev= install_type= isdir=false stripme= no_mode=: for arg do arg2= if test -n "$dest"; then func_append files " $dest" dest=$arg continue fi case $arg in -d) isdir=: ;; -f) if $install_cp; then :; else prev=$arg fi ;; -g | -m | -o) prev=$arg ;; -s) stripme=" -s" continue ;; -*) ;; *) # If the previous option needed an argument, then skip it. if test -n "$prev"; then if test X-m = "X$prev" && test -n "$install_override_mode"; then arg2=$install_override_mode no_mode=false fi prev= else dest=$arg continue fi ;; esac # Aesthetically quote the argument. func_quote_for_eval "$arg" func_append install_prog " $func_quote_for_eval_result" if test -n "$arg2"; then func_quote_for_eval "$arg2" fi func_append install_shared_prog " $func_quote_for_eval_result" done test -z "$install_prog" && \ func_fatal_help "you must specify an install program" test -n "$prev" && \ func_fatal_help "the '$prev' option requires an argument" if test -n "$install_override_mode" && $no_mode; then if $install_cp; then :; else func_quote_for_eval "$install_override_mode" func_append install_shared_prog " -m $func_quote_for_eval_result" fi fi if test -z "$files"; then if test -z "$dest"; then func_fatal_help "no file or destination specified" else func_fatal_help "you must specify a destination" fi fi # Strip any trailing slash from the destination. func_stripname '' '/' "$dest" dest=$func_stripname_result # Check to see that the destination is a directory. test -d "$dest" && isdir=: if $isdir; then destdir=$dest destname= else func_dirname_and_basename "$dest" "" "." destdir=$func_dirname_result destname=$func_basename_result # Not a directory, so check to see that there is only one file specified. set dummy $files; shift test "$#" -gt 1 && \ func_fatal_help "'$dest' is not a directory" fi case $destdir in [\\/]* | [A-Za-z]:[\\/]*) ;; *) for file in $files; do case $file in *.lo) ;; *) func_fatal_help "'$destdir' must be an absolute directory name" ;; esac done ;; esac # This variable tells wrapper scripts just to set variables rather # than running their programs. libtool_install_magic=$magic staticlibs= future_libdirs= current_libdirs= for file in $files; do # Do each installation. case $file in *.$libext) # Do the static libraries later. func_append staticlibs " $file" ;; *.la) func_resolve_sysroot "$file" file=$func_resolve_sysroot_result # Check to see that this really is a libtool archive. func_lalib_unsafe_p "$file" \ || func_fatal_help "'$file' is not a valid libtool archive" library_names= old_library= relink_command= func_source "$file" # Add the libdir to current_libdirs if it is the destination. if test "X$destdir" = "X$libdir"; then case "$current_libdirs " in *" $libdir "*) ;; *) func_append current_libdirs " $libdir" ;; esac else # Note the libdir as a future libdir. case "$future_libdirs " in *" $libdir "*) ;; *) func_append future_libdirs " $libdir" ;; esac fi func_dirname "$file" "/" "" dir=$func_dirname_result func_append dir "$objdir" if test -n "$relink_command"; then # Determine the prefix the user has applied to our future dir. inst_prefix_dir=`$ECHO "$destdir" | $SED -e "s%$libdir\$%%"` # Don't allow the user to place us outside of our expected # location b/c this prevents finding dependent libraries that # are installed to the same prefix. # At present, this check doesn't affect windows .dll's that # are installed into $libdir/../bin (currently, that works fine) # but it's something to keep an eye on. test "$inst_prefix_dir" = "$destdir" && \ func_fatal_error "error: cannot install '$file' to a directory not ending in $libdir" if test -n "$inst_prefix_dir"; then # Stick the inst_prefix_dir data into the link command. relink_command=`$ECHO "$relink_command" | $SED "s%@inst_prefix_dir@%-inst-prefix-dir $inst_prefix_dir%"` else relink_command=`$ECHO "$relink_command" | $SED "s%@inst_prefix_dir@%%"` fi func_warning "relinking '$file'" func_show_eval "$relink_command" \ 'func_fatal_error "error: relink '\''$file'\'' with the above command before installing it"' fi # See the names of the shared library. set dummy $library_names; shift if test -n "$1"; then realname=$1 shift srcname=$realname test -n "$relink_command" && srcname=${realname}T # Install the shared library and build the symlinks. func_show_eval "$install_shared_prog $dir/$srcname $destdir/$realname" \ 'exit $?' tstripme=$stripme case $host_os in cygwin* | mingw* | pw32* | cegcc*) case $realname in *.dll.a) tstripme= ;; esac ;; os2*) case $realname in *_dll.a) tstripme= ;; esac ;; esac if test -n "$tstripme" && test -n "$striplib"; then func_show_eval "$striplib $destdir/$realname" 'exit $?' fi if test "$#" -gt 0; then # Delete the old symlinks, and create new ones. # Try 'ln -sf' first, because the 'ln' binary might depend on # the symlink we replace! Solaris /bin/ln does not understand -f, # so we also need to try rm && ln -s. for linkname do test "$linkname" != "$realname" \ && func_show_eval "(cd $destdir && { $LN_S -f $realname $linkname || { $RM $linkname && $LN_S $realname $linkname; }; })" done fi # Do each command in the postinstall commands. lib=$destdir/$realname func_execute_cmds "$postinstall_cmds" 'exit $?' fi # Install the pseudo-library for information purposes. func_basename "$file" name=$func_basename_result instname=$dir/${name}i func_show_eval "$install_prog $instname $destdir/$name" 'exit $?' # Maybe install the static library, too. test -n "$old_library" && func_append staticlibs " $dir/$old_library" ;; *.lo) # Install (i.e. copy) a libtool object. # Figure out destination file name, if it wasn't already specified. if test -n "$destname"; then destfile=$destdir/$destname else func_basename "$file" destfile=$func_basename_result destfile=$destdir/$destfile fi # Deduce the name of the destination old-style object file. case $destfile in *.lo) func_lo2o "$destfile" staticdest=$func_lo2o_result ;; *.$objext) staticdest=$destfile destfile= ;; *) func_fatal_help "cannot copy a libtool object to '$destfile'" ;; esac # Install the libtool object if requested. test -n "$destfile" && \ func_show_eval "$install_prog $file $destfile" 'exit $?' # Install the old object if enabled. if test yes = "$build_old_libs"; then # Deduce the name of the old-style object file. func_lo2o "$file" staticobj=$func_lo2o_result func_show_eval "$install_prog \$staticobj \$staticdest" 'exit $?' fi exit $EXIT_SUCCESS ;; *) # Figure out destination file name, if it wasn't already specified. if test -n "$destname"; then destfile=$destdir/$destname else func_basename "$file" destfile=$func_basename_result destfile=$destdir/$destfile fi # If the file is missing, and there is a .exe on the end, strip it # because it is most likely a libtool script we actually want to # install stripped_ext= case $file in *.exe) if test ! -f "$file"; then func_stripname '' '.exe' "$file" file=$func_stripname_result stripped_ext=.exe fi ;; esac # Do a test to see if this is really a libtool program. case $host in *cygwin* | *mingw*) if func_ltwrapper_executable_p "$file"; then func_ltwrapper_scriptname "$file" wrapper=$func_ltwrapper_scriptname_result else func_stripname '' '.exe' "$file" wrapper=$func_stripname_result fi ;; *) wrapper=$file ;; esac if func_ltwrapper_script_p "$wrapper"; then notinst_deplibs= relink_command= func_source "$wrapper" # Check the variables that should have been set. test -z "$generated_by_libtool_version" && \ func_fatal_error "invalid libtool wrapper script '$wrapper'" finalize=: for lib in $notinst_deplibs; do # Check to see that each library is installed. libdir= if test -f "$lib"; then func_source "$lib" fi libfile=$libdir/`$ECHO "$lib" | $SED 's%^.*/%%g'` if test -n "$libdir" && test ! -f "$libfile"; then func_warning "'$lib' has not been installed in '$libdir'" finalize=false fi done relink_command= func_source "$wrapper" outputname= if test no = "$fast_install" && test -n "$relink_command"; then $opt_dry_run || { if $finalize; then tmpdir=`func_mktempdir` func_basename "$file$stripped_ext" file=$func_basename_result outputname=$tmpdir/$file # Replace the output file specification. relink_command=`$ECHO "$relink_command" | $SED 's%@OUTPUT@%'"$outputname"'%g'` $opt_quiet || { func_quote_for_expand "$relink_command" eval "func_echo $func_quote_for_expand_result" } if eval "$relink_command"; then : else func_error "error: relink '$file' with the above command before installing it" $opt_dry_run || ${RM}r "$tmpdir" continue fi file=$outputname else func_warning "cannot relink '$file'" fi } else # Install the binary that we compiled earlier. file=`$ECHO "$file$stripped_ext" | $SED "s%\([^/]*\)$%$objdir/\1%"` fi fi # remove .exe since cygwin /usr/bin/install will append another # one anyway case $install_prog,$host in */usr/bin/install*,*cygwin*) case $file:$destfile in *.exe:*.exe) # this is ok ;; *.exe:*) destfile=$destfile.exe ;; *:*.exe) func_stripname '' '.exe' "$destfile" destfile=$func_stripname_result ;; esac ;; esac func_show_eval "$install_prog\$stripme \$file \$destfile" 'exit $?' $opt_dry_run || if test -n "$outputname"; then ${RM}r "$tmpdir" fi ;; esac done for file in $staticlibs; do func_basename "$file" name=$func_basename_result # Set up the ranlib parameters. oldlib=$destdir/$name func_to_tool_file "$oldlib" func_convert_file_msys_to_w32 tool_oldlib=$func_to_tool_file_result func_show_eval "$install_prog \$file \$oldlib" 'exit $?' if test -n "$stripme" && test -n "$old_striplib"; then func_show_eval "$old_striplib $tool_oldlib" 'exit $?' fi # Do each command in the postinstall commands. func_execute_cmds "$old_postinstall_cmds" 'exit $?' done test -n "$future_libdirs" && \ func_warning "remember to run '$progname --finish$future_libdirs'" if test -n "$current_libdirs"; then # Maybe just do a dry run. $opt_dry_run && current_libdirs=" -n$current_libdirs" exec_cmd='$SHELL "$progpath" $preserve_args --finish$current_libdirs' else exit $EXIT_SUCCESS fi } test install = "$opt_mode" && func_mode_install ${1+"$@"} # func_generate_dlsyms outputname originator pic_p # Extract symbols from dlprefiles and create ${outputname}S.o with # a dlpreopen symbol table. func_generate_dlsyms () { $debug_cmd my_outputname=$1 my_originator=$2 my_pic_p=${3-false} my_prefix=`$ECHO "$my_originator" | $SED 's%[^a-zA-Z0-9]%_%g'` my_dlsyms= if test -n "$dlfiles$dlprefiles" || test no != "$dlself"; then if test -n "$NM" && test -n "$global_symbol_pipe"; then my_dlsyms=${my_outputname}S.c else func_error "not configured to extract global symbols from dlpreopened files" fi fi if test -n "$my_dlsyms"; then case $my_dlsyms in "") ;; *.c) # Discover the nlist of each of the dlfiles. nlist=$output_objdir/$my_outputname.nm func_show_eval "$RM $nlist ${nlist}S ${nlist}T" # Parse the name list into a source file. func_verbose "creating $output_objdir/$my_dlsyms" $opt_dry_run || $ECHO > "$output_objdir/$my_dlsyms" "\ /* $my_dlsyms - symbol resolution table for '$my_outputname' dlsym emulation. */ /* Generated by $PROGRAM (GNU $PACKAGE) $VERSION */ #ifdef __cplusplus extern \"C\" { #endif #if defined __GNUC__ && (((__GNUC__ == 4) && (__GNUC_MINOR__ >= 4)) || (__GNUC__ > 4)) #pragma GCC diagnostic ignored \"-Wstrict-prototypes\" #endif /* Keep this code in sync between libtool.m4, ltmain, lt_system.h, and tests. */ #if defined _WIN32 || defined __CYGWIN__ || defined _WIN32_WCE /* DATA imports from DLLs on WIN32 can't be const, because runtime relocations are performed -- see ld's documentation on pseudo-relocs. */ # define LT_DLSYM_CONST #elif defined __osf__ /* This system does not cope well with relocations in const data. */ # define LT_DLSYM_CONST #else # define LT_DLSYM_CONST const #endif #define STREQ(s1, s2) (strcmp ((s1), (s2)) == 0) /* External symbol declarations for the compiler. */\ " if test yes = "$dlself"; then func_verbose "generating symbol list for '$output'" $opt_dry_run || echo ': @PROGRAM@ ' > "$nlist" # Add our own program objects to the symbol list. progfiles=`$ECHO "$objs$old_deplibs" | $SP2NL | $SED "$lo2o" | $NL2SP` for progfile in $progfiles; do func_to_tool_file "$progfile" func_convert_file_msys_to_w32 func_verbose "extracting global C symbols from '$func_to_tool_file_result'" $opt_dry_run || eval "$NM $func_to_tool_file_result | $global_symbol_pipe >> '$nlist'" done if test -n "$exclude_expsyms"; then $opt_dry_run || { eval '$EGREP -v " ($exclude_expsyms)$" "$nlist" > "$nlist"T' eval '$MV "$nlist"T "$nlist"' } fi if test -n "$export_symbols_regex"; then $opt_dry_run || { eval '$EGREP -e "$export_symbols_regex" "$nlist" > "$nlist"T' eval '$MV "$nlist"T "$nlist"' } fi # Prepare the list of exported symbols if test -z "$export_symbols"; then export_symbols=$output_objdir/$outputname.exp $opt_dry_run || { $RM $export_symbols eval "$SED -n -e '/^: @PROGRAM@ $/d' -e 's/^.* \(.*\)$/\1/p' "'< "$nlist" > "$export_symbols"' case $host in *cygwin* | *mingw* | *cegcc* ) eval "echo EXPORTS "'> "$output_objdir/$outputname.def"' eval 'cat "$export_symbols" >> "$output_objdir/$outputname.def"' ;; esac } else $opt_dry_run || { eval "$SED -e 's/\([].[*^$]\)/\\\\\1/g' -e 's/^/ /' -e 's/$/$/'"' < "$export_symbols" > "$output_objdir/$outputname.exp"' eval '$GREP -f "$output_objdir/$outputname.exp" < "$nlist" > "$nlist"T' eval '$MV "$nlist"T "$nlist"' case $host in *cygwin* | *mingw* | *cegcc* ) eval "echo EXPORTS "'> "$output_objdir/$outputname.def"' eval 'cat "$nlist" >> "$output_objdir/$outputname.def"' ;; esac } fi fi for dlprefile in $dlprefiles; do func_verbose "extracting global C symbols from '$dlprefile'" func_basename "$dlprefile" name=$func_basename_result case $host in *cygwin* | *mingw* | *cegcc* ) # if an import library, we need to obtain dlname if func_win32_import_lib_p "$dlprefile"; then func_tr_sh "$dlprefile" eval "curr_lafile=\$libfile_$func_tr_sh_result" dlprefile_dlbasename= if test -n "$curr_lafile" && func_lalib_p "$curr_lafile"; then # Use subshell, to avoid clobbering current variable values dlprefile_dlname=`source "$curr_lafile" && echo "$dlname"` if test -n "$dlprefile_dlname"; then func_basename "$dlprefile_dlname" dlprefile_dlbasename=$func_basename_result else # no lafile. user explicitly requested -dlpreopen <import library>. $sharedlib_from_linklib_cmd "$dlprefile" dlprefile_dlbasename=$sharedlib_from_linklib_result fi fi $opt_dry_run || { if test -n "$dlprefile_dlbasename"; then eval '$ECHO ": $dlprefile_dlbasename" >> "$nlist"' else func_warning "Could not compute DLL name from $name" eval '$ECHO ": $name " >> "$nlist"' fi func_to_tool_file "$dlprefile" func_convert_file_msys_to_w32 eval "$NM \"$func_to_tool_file_result\" 2>/dev/null | $global_symbol_pipe | $SED -e '/I __imp/d' -e 's/I __nm_/D /;s/_nm__//' >> '$nlist'" } else # not an import lib $opt_dry_run || { eval '$ECHO ": $name " >> "$nlist"' func_to_tool_file "$dlprefile" func_convert_file_msys_to_w32 eval "$NM \"$func_to_tool_file_result\" 2>/dev/null | $global_symbol_pipe >> '$nlist'" } fi ;; *) $opt_dry_run || { eval '$ECHO ": $name " >> "$nlist"' func_to_tool_file "$dlprefile" func_convert_file_msys_to_w32 eval "$NM \"$func_to_tool_file_result\" 2>/dev/null | $global_symbol_pipe >> '$nlist'" } ;; esac done $opt_dry_run || { # Make sure we have at least an empty file. test -f "$nlist" || : > "$nlist" if test -n "$exclude_expsyms"; then $EGREP -v " ($exclude_expsyms)$" "$nlist" > "$nlist"T $MV "$nlist"T "$nlist" fi # Try sorting and uniquifying the output. if $GREP -v "^: " < "$nlist" | if sort -k 3 </dev/null >/dev/null 2>&1; then sort -k 3 else sort +2 fi | uniq > "$nlist"S; then : else $GREP -v "^: " < "$nlist" > "$nlist"S fi if test -f "$nlist"S; then eval "$global_symbol_to_cdecl"' < "$nlist"S >> "$output_objdir/$my_dlsyms"' else echo '/* NONE */' >> "$output_objdir/$my_dlsyms" fi func_show_eval '$RM "${nlist}I"' if test -n "$global_symbol_to_import"; then eval "$global_symbol_to_import"' < "$nlist"S > "$nlist"I' fi echo >> "$output_objdir/$my_dlsyms" "\ /* The mapping between symbol names and symbols. */ typedef struct { const char *name; void *address; } lt_dlsymlist; extern LT_DLSYM_CONST lt_dlsymlist lt_${my_prefix}_LTX_preloaded_symbols[];\ " if test -s "$nlist"I; then echo >> "$output_objdir/$my_dlsyms" "\ static void lt_syminit(void) { LT_DLSYM_CONST lt_dlsymlist *symbol = lt_${my_prefix}_LTX_preloaded_symbols; for (; symbol->name; ++symbol) {" $SED 's/.*/ if (STREQ (symbol->name, \"&\")) symbol->address = (void *) \&&;/' < "$nlist"I >> "$output_objdir/$my_dlsyms" echo >> "$output_objdir/$my_dlsyms" "\ } }" fi echo >> "$output_objdir/$my_dlsyms" "\ LT_DLSYM_CONST lt_dlsymlist lt_${my_prefix}_LTX_preloaded_symbols[] = { {\"$my_originator\", (void *) 0}," if test -s "$nlist"I; then echo >> "$output_objdir/$my_dlsyms" "\ {\"@INIT@\", (void *) &lt_syminit}," fi case $need_lib_prefix in no) eval "$global_symbol_to_c_name_address" < "$nlist" >> "$output_objdir/$my_dlsyms" ;; *) eval "$global_symbol_to_c_name_address_lib_prefix" < "$nlist" >> "$output_objdir/$my_dlsyms" ;; esac echo >> "$output_objdir/$my_dlsyms" "\ {0, (void *) 0} }; /* This works around a problem in FreeBSD linker */ #ifdef FREEBSD_WORKAROUND static const void *lt_preloaded_setup() { return lt_${my_prefix}_LTX_preloaded_symbols; } #endif #ifdef __cplusplus } #endif\ " } # !$opt_dry_run pic_flag_for_symtable= case "$compile_command " in *" -static "*) ;; *) case $host in # compiling the symbol table file with pic_flag works around # a FreeBSD bug that causes programs to crash when -lm is # linked before any other PIC object. But we must not use # pic_flag when linking with -static. The problem exists in # FreeBSD 2.2.6 and is fixed in FreeBSD 3.1. *-*-freebsd2.*|*-*-freebsd3.0*|*-*-freebsdelf3.0*) pic_flag_for_symtable=" $pic_flag -DFREEBSD_WORKAROUND" ;; *-*-hpux*) pic_flag_for_symtable=" $pic_flag" ;; *) $my_pic_p && pic_flag_for_symtable=" $pic_flag" ;; esac ;; esac symtab_cflags= for arg in $LTCFLAGS; do case $arg in -pie | -fpie | -fPIE) ;; *) func_append symtab_cflags " $arg" ;; esac done # Now compile the dynamic symbol file. func_show_eval '(cd $output_objdir && $LTCC$symtab_cflags -c$no_builtin_flag$pic_flag_for_symtable "$my_dlsyms")' 'exit $?' # Clean up the generated files. func_show_eval '$RM "$output_objdir/$my_dlsyms" "$nlist" "${nlist}S" "${nlist}T" "${nlist}I"' # Transform the symbol file into the correct name. symfileobj=$output_objdir/${my_outputname}S.$objext case $host in *cygwin* | *mingw* | *cegcc* ) if test -f "$output_objdir/$my_outputname.def"; then compile_command=`$ECHO "$compile_command" | $SED "s%@SYMFILE@%$output_objdir/$my_outputname.def $symfileobj%"` finalize_command=`$ECHO "$finalize_command" | $SED "s%@SYMFILE@%$output_objdir/$my_outputname.def $symfileobj%"` else compile_command=`$ECHO "$compile_command" | $SED "s%@SYMFILE@%$symfileobj%"` finalize_command=`$ECHO "$finalize_command" | $SED "s%@SYMFILE@%$symfileobj%"` fi ;; *) compile_command=`$ECHO "$compile_command" | $SED "s%@SYMFILE@%$symfileobj%"` finalize_command=`$ECHO "$finalize_command" | $SED "s%@SYMFILE@%$symfileobj%"` ;; esac ;; *) func_fatal_error "unknown suffix for '$my_dlsyms'" ;; esac else # We keep going just in case the user didn't refer to # lt_preloaded_symbols. The linker will fail if global_symbol_pipe # really was required. # Nullify the symbol file. compile_command=`$ECHO "$compile_command" | $SED "s% @SYMFILE@%%"` finalize_command=`$ECHO "$finalize_command" | $SED "s% @SYMFILE@%%"` fi } # func_cygming_gnu_implib_p ARG # This predicate returns with zero status (TRUE) if # ARG is a GNU/binutils-style import library. Returns # with nonzero status (FALSE) otherwise. func_cygming_gnu_implib_p () { $debug_cmd func_to_tool_file "$1" func_convert_file_msys_to_w32 func_cygming_gnu_implib_tmp=`$NM "$func_to_tool_file_result" | eval "$global_symbol_pipe" | $EGREP ' (_head_[A-Za-z0-9_]+_[ad]l*|[A-Za-z0-9_]+_[ad]l*_iname)$'` test -n "$func_cygming_gnu_implib_tmp" } # func_cygming_ms_implib_p ARG # This predicate returns with zero status (TRUE) if # ARG is an MS-style import library. Returns # with nonzero status (FALSE) otherwise. func_cygming_ms_implib_p () { $debug_cmd func_to_tool_file "$1" func_convert_file_msys_to_w32 func_cygming_ms_implib_tmp=`$NM "$func_to_tool_file_result" | eval "$global_symbol_pipe" | $GREP '_NULL_IMPORT_DESCRIPTOR'` test -n "$func_cygming_ms_implib_tmp" } # func_win32_libid arg # return the library type of file 'arg' # # Need a lot of goo to handle *both* DLLs and import libs # Has to be a shell function in order to 'eat' the argument # that is supplied when $file_magic_command is called. # Despite the name, also deal with 64 bit binaries. func_win32_libid () { $debug_cmd win32_libid_type=unknown win32_fileres=`file -L $1 2>/dev/null` case $win32_fileres in *ar\ archive\ import\ library*) # definitely import win32_libid_type="x86 archive import" ;; *ar\ archive*) # could be an import, or static # Keep the egrep pattern in sync with the one in _LT_CHECK_MAGIC_METHOD. if eval $OBJDUMP -f $1 | $SED -e '10q' 2>/dev/null | $EGREP 'file format (pei*-i386(.*architecture: i386)?|pe-arm-wince|pe-x86-64)' >/dev/null; then case $nm_interface in "MS dumpbin") if func_cygming_ms_implib_p "$1" || func_cygming_gnu_implib_p "$1" then win32_nmres=import else win32_nmres= fi ;; *) func_to_tool_file "$1" func_convert_file_msys_to_w32 win32_nmres=`eval $NM -f posix -A \"$func_to_tool_file_result\" | $SED -n -e ' 1,100{ / I /{ s|.*|import| p q } }'` ;; esac case $win32_nmres in import*) win32_libid_type="x86 archive import";; *) win32_libid_type="x86 archive static";; esac fi ;; *DLL*) win32_libid_type="x86 DLL" ;; *executable*) # but shell scripts are "executable" too... case $win32_fileres in *MS\ Windows\ PE\ Intel*) win32_libid_type="x86 DLL" ;; esac ;; esac $ECHO "$win32_libid_type" } # func_cygming_dll_for_implib ARG # # Platform-specific function to extract the # name of the DLL associated with the specified # import library ARG. # Invoked by eval'ing the libtool variable # $sharedlib_from_linklib_cmd # Result is available in the variable # $sharedlib_from_linklib_result func_cygming_dll_for_implib () { $debug_cmd sharedlib_from_linklib_result=`$DLLTOOL --identify-strict --identify "$1"` } # func_cygming_dll_for_implib_fallback_core SECTION_NAME LIBNAMEs # # The is the core of a fallback implementation of a # platform-specific function to extract the name of the # DLL associated with the specified import library LIBNAME. # # SECTION_NAME is either .idata$6 or .idata$7, depending # on the platform and compiler that created the implib. # # Echos the name of the DLL associated with the # specified import library. func_cygming_dll_for_implib_fallback_core () { $debug_cmd match_literal=`$ECHO "$1" | $SED "$sed_make_literal_regex"` $OBJDUMP -s --section "$1" "$2" 2>/dev/null | $SED '/^Contents of section '"$match_literal"':/{ # Place marker at beginning of archive member dllname section s/.*/====MARK====/ p d } # These lines can sometimes be longer than 43 characters, but # are always uninteresting /:[ ]*file format pe[i]\{,1\}-/d /^In archive [^:]*:/d # Ensure marker is printed /^====MARK====/p # Remove all lines with less than 43 characters /^.\{43\}/!d # From remaining lines, remove first 43 characters s/^.\{43\}//' | $SED -n ' # Join marker and all lines until next marker into a single line /^====MARK====/ b para H $ b para b :para x s/\n//g # Remove the marker s/^====MARK====// # Remove trailing dots and whitespace s/[\. \t]*$// # Print /./p' | # we now have a list, one entry per line, of the stringified # contents of the appropriate section of all members of the # archive that possess that section. Heuristic: eliminate # all those that have a first or second character that is # a '.' (that is, objdump's representation of an unprintable # character.) This should work for all archives with less than # 0x302f exports -- but will fail for DLLs whose name actually # begins with a literal '.' or a single character followed by # a '.'. # # Of those that remain, print the first one. $SED -e '/^\./d;/^.\./d;q' } # func_cygming_dll_for_implib_fallback ARG # Platform-specific function to extract the # name of the DLL associated with the specified # import library ARG. # # This fallback implementation is for use when $DLLTOOL # does not support the --identify-strict option. # Invoked by eval'ing the libtool variable # $sharedlib_from_linklib_cmd # Result is available in the variable # $sharedlib_from_linklib_result func_cygming_dll_for_implib_fallback () { $debug_cmd if func_cygming_gnu_implib_p "$1"; then # binutils import library sharedlib_from_linklib_result=`func_cygming_dll_for_implib_fallback_core '.idata$7' "$1"` elif func_cygming_ms_implib_p "$1"; then # ms-generated import library sharedlib_from_linklib_result=`func_cygming_dll_for_implib_fallback_core '.idata$6' "$1"` else # unknown sharedlib_from_linklib_result= fi } # func_extract_an_archive dir oldlib func_extract_an_archive () { $debug_cmd f_ex_an_ar_dir=$1; shift f_ex_an_ar_oldlib=$1 if test yes = "$lock_old_archive_extraction"; then lockfile=$f_ex_an_ar_oldlib.lock until $opt_dry_run || ln "$progpath" "$lockfile" 2>/dev/null; do func_echo "Waiting for $lockfile to be removed" sleep 2 done fi func_show_eval "(cd \$f_ex_an_ar_dir && $AR x \"\$f_ex_an_ar_oldlib\")" \ 'stat=$?; rm -f "$lockfile"; exit $stat' if test yes = "$lock_old_archive_extraction"; then $opt_dry_run || rm -f "$lockfile" fi if ($AR t "$f_ex_an_ar_oldlib" | sort | sort -uc >/dev/null 2>&1); then : else func_fatal_error "object name conflicts in archive: $f_ex_an_ar_dir/$f_ex_an_ar_oldlib" fi } # func_extract_archives gentop oldlib ... func_extract_archives () { $debug_cmd my_gentop=$1; shift my_oldlibs=${1+"$@"} my_oldobjs= my_xlib= my_xabs= my_xdir= for my_xlib in $my_oldlibs; do # Extract the objects. case $my_xlib in [\\/]* | [A-Za-z]:[\\/]*) my_xabs=$my_xlib ;; *) my_xabs=`pwd`"/$my_xlib" ;; esac func_basename "$my_xlib" my_xlib=$func_basename_result my_xlib_u=$my_xlib while :; do case " $extracted_archives " in *" $my_xlib_u "*) func_arith $extracted_serial + 1 extracted_serial=$func_arith_result my_xlib_u=lt$extracted_serial-$my_xlib ;; *) break ;; esac done extracted_archives="$extracted_archives $my_xlib_u" my_xdir=$my_gentop/$my_xlib_u func_mkdir_p "$my_xdir" case $host in *-darwin*) func_verbose "Extracting $my_xabs" # Do not bother doing anything if just a dry run $opt_dry_run || { darwin_orig_dir=`pwd` cd $my_xdir || exit $? darwin_archive=$my_xabs darwin_curdir=`pwd` func_basename "$darwin_archive" darwin_base_archive=$func_basename_result darwin_arches=`$LIPO -info "$darwin_archive" 2>/dev/null | $GREP Architectures 2>/dev/null || true` if test -n "$darwin_arches"; then darwin_arches=`$ECHO "$darwin_arches" | $SED -e 's/.*are://'` darwin_arch= func_verbose "$darwin_base_archive has multiple architectures $darwin_arches" for darwin_arch in $darwin_arches; do func_mkdir_p "unfat-$$/$darwin_base_archive-$darwin_arch" $LIPO -thin $darwin_arch -output "unfat-$$/$darwin_base_archive-$darwin_arch/$darwin_base_archive" "$darwin_archive" cd "unfat-$$/$darwin_base_archive-$darwin_arch" func_extract_an_archive "`pwd`" "$darwin_base_archive" cd "$darwin_curdir" $RM "unfat-$$/$darwin_base_archive-$darwin_arch/$darwin_base_archive" done # $darwin_arches ## Okay now we've a bunch of thin objects, gotta fatten them up :) darwin_filelist=`find unfat-$$ -type f -name \*.o -print -o -name \*.lo -print | $SED -e "$sed_basename" | sort -u` darwin_file= darwin_files= for darwin_file in $darwin_filelist; do darwin_files=`find unfat-$$ -name $darwin_file -print | sort | $NL2SP` $LIPO -create -output "$darwin_file" $darwin_files done # $darwin_filelist $RM -rf unfat-$$ cd "$darwin_orig_dir" else cd $darwin_orig_dir func_extract_an_archive "$my_xdir" "$my_xabs" fi # $darwin_arches } # !$opt_dry_run ;; *) func_extract_an_archive "$my_xdir" "$my_xabs" ;; esac my_oldobjs="$my_oldobjs "`find $my_xdir -name \*.$objext -print -o -name \*.lo -print | sort | $NL2SP` done func_extract_archives_result=$my_oldobjs } # func_emit_wrapper [arg=no] # # Emit a libtool wrapper script on stdout. # Don't directly open a file because we may want to # incorporate the script contents within a cygwin/mingw # wrapper executable. Must ONLY be called from within # func_mode_link because it depends on a number of variables # set therein. # # ARG is the value that the WRAPPER_SCRIPT_BELONGS_IN_OBJDIR # variable will take. If 'yes', then the emitted script # will assume that the directory where it is stored is # the $objdir directory. This is a cygwin/mingw-specific # behavior. func_emit_wrapper () { func_emit_wrapper_arg1=${1-no} $ECHO "\ #! $SHELL # $output - temporary wrapper script for $objdir/$outputname # Generated by $PROGRAM (GNU $PACKAGE) $VERSION # # The $output program cannot be directly executed until all the libtool # libraries that it depends on are installed. # # This wrapper script should never be moved out of the build directory. # If it is, it will not operate correctly. # Sed substitution that helps us do robust quoting. It backslashifies # metacharacters that are still active within double-quoted strings. sed_quote_subst='$sed_quote_subst' # Be Bourne compatible if test -n \"\${ZSH_VERSION+set}\" && (emulate sh) >/dev/null 2>&1; then emulate sh NULLCMD=: # Zsh 3.x and 4.x performs word splitting on \${1+\"\$@\"}, which # is contrary to our usage. Disable this feature. alias -g '\${1+\"\$@\"}'='\"\$@\"' setopt NO_GLOB_SUBST else case \`(set -o) 2>/dev/null\` in *posix*) set -o posix;; esac fi BIN_SH=xpg4; export BIN_SH # for Tru64 DUALCASE=1; export DUALCASE # for MKS sh # The HP-UX ksh and POSIX shell print the target directory to stdout # if CDPATH is set. (unset CDPATH) >/dev/null 2>&1 && unset CDPATH relink_command=\"$relink_command\" # This environment variable determines our operation mode. if test \"\$libtool_install_magic\" = \"$magic\"; then # install mode needs the following variables: generated_by_libtool_version='$macro_version' notinst_deplibs='$notinst_deplibs' else # When we are sourced in execute mode, \$file and \$ECHO are already set. if test \"\$libtool_execute_magic\" != \"$magic\"; then file=\"\$0\"" qECHO=`$ECHO "$ECHO" | $SED "$sed_quote_subst"` $ECHO "\ # A function that is used when there is no print builtin or printf. func_fallback_echo () { eval 'cat <<_LTECHO_EOF \$1 _LTECHO_EOF' } ECHO=\"$qECHO\" fi # Very basic option parsing. These options are (a) specific to # the libtool wrapper, (b) are identical between the wrapper # /script/ and the wrapper /executable/ that is used only on # windows platforms, and (c) all begin with the string "--lt-" # (application programs are unlikely to have options that match # this pattern). # # There are only two supported options: --lt-debug and # --lt-dump-script. There is, deliberately, no --lt-help. # # The first argument to this parsing function should be the # script's $0 value, followed by "$@". lt_option_debug= func_parse_lt_options () { lt_script_arg0=\$0 shift for lt_opt do case \"\$lt_opt\" in --lt-debug) lt_option_debug=1 ;; --lt-dump-script) lt_dump_D=\`\$ECHO \"X\$lt_script_arg0\" | $SED -e 's/^X//' -e 's%/[^/]*$%%'\` test \"X\$lt_dump_D\" = \"X\$lt_script_arg0\" && lt_dump_D=. lt_dump_F=\`\$ECHO \"X\$lt_script_arg0\" | $SED -e 's/^X//' -e 's%^.*/%%'\` cat \"\$lt_dump_D/\$lt_dump_F\" exit 0 ;; --lt-*) \$ECHO \"Unrecognized --lt- option: '\$lt_opt'\" 1>&2 exit 1 ;; esac done # Print the debug banner immediately: if test -n \"\$lt_option_debug\"; then echo \"$outputname:$output:\$LINENO: libtool wrapper (GNU $PACKAGE) $VERSION\" 1>&2 fi } # Used when --lt-debug. Prints its arguments to stdout # (redirection is the responsibility of the caller) func_lt_dump_args () { lt_dump_args_N=1; for lt_arg do \$ECHO \"$outputname:$output:\$LINENO: newargv[\$lt_dump_args_N]: \$lt_arg\" lt_dump_args_N=\`expr \$lt_dump_args_N + 1\` done } # Core function for launching the target application func_exec_program_core () { " case $host in # Backslashes separate directories on plain windows *-*-mingw | *-*-os2* | *-cegcc*) $ECHO "\ if test -n \"\$lt_option_debug\"; then \$ECHO \"$outputname:$output:\$LINENO: newargv[0]: \$progdir\\\\\$program\" 1>&2 func_lt_dump_args \${1+\"\$@\"} 1>&2 fi exec \"\$progdir\\\\\$program\" \${1+\"\$@\"} " ;; *) $ECHO "\ if test -n \"\$lt_option_debug\"; then \$ECHO \"$outputname:$output:\$LINENO: newargv[0]: \$progdir/\$program\" 1>&2 func_lt_dump_args \${1+\"\$@\"} 1>&2 fi exec \"\$progdir/\$program\" \${1+\"\$@\"} " ;; esac $ECHO "\ \$ECHO \"\$0: cannot exec \$program \$*\" 1>&2 exit 1 } # A function to encapsulate launching the target application # Strips options in the --lt-* namespace from \$@ and # launches target application with the remaining arguments. func_exec_program () { case \" \$* \" in *\\ --lt-*) for lt_wr_arg do case \$lt_wr_arg in --lt-*) ;; *) set x \"\$@\" \"\$lt_wr_arg\"; shift;; esac shift done ;; esac func_exec_program_core \${1+\"\$@\"} } # Parse options func_parse_lt_options \"\$0\" \${1+\"\$@\"} # Find the directory that this script lives in. thisdir=\`\$ECHO \"\$file\" | $SED 's%/[^/]*$%%'\` test \"x\$thisdir\" = \"x\$file\" && thisdir=. # Follow symbolic links until we get to the real thisdir. file=\`ls -ld \"\$file\" | $SED -n 's/.*-> //p'\` while test -n \"\$file\"; do destdir=\`\$ECHO \"\$file\" | $SED 's%/[^/]*\$%%'\` # If there was a directory component, then change thisdir. if test \"x\$destdir\" != \"x\$file\"; then case \"\$destdir\" in [\\\\/]* | [A-Za-z]:[\\\\/]*) thisdir=\"\$destdir\" ;; *) thisdir=\"\$thisdir/\$destdir\" ;; esac fi file=\`\$ECHO \"\$file\" | $SED 's%^.*/%%'\` file=\`ls -ld \"\$thisdir/\$file\" | $SED -n 's/.*-> //p'\` done # Usually 'no', except on cygwin/mingw when embedded into # the cwrapper. WRAPPER_SCRIPT_BELONGS_IN_OBJDIR=$func_emit_wrapper_arg1 if test \"\$WRAPPER_SCRIPT_BELONGS_IN_OBJDIR\" = \"yes\"; then # special case for '.' if test \"\$thisdir\" = \".\"; then thisdir=\`pwd\` fi # remove .libs from thisdir case \"\$thisdir\" in *[\\\\/]$objdir ) thisdir=\`\$ECHO \"\$thisdir\" | $SED 's%[\\\\/][^\\\\/]*$%%'\` ;; $objdir ) thisdir=. ;; esac fi # Try to get the absolute directory name. absdir=\`cd \"\$thisdir\" && pwd\` test -n \"\$absdir\" && thisdir=\"\$absdir\" " if test yes = "$fast_install"; then $ECHO "\ program=lt-'$outputname'$exeext progdir=\"\$thisdir/$objdir\" if test ! -f \"\$progdir/\$program\" || { file=\`ls -1dt \"\$progdir/\$program\" \"\$progdir/../\$program\" 2>/dev/null | $SED 1q\`; \\ test \"X\$file\" != \"X\$progdir/\$program\"; }; then file=\"\$\$-\$program\" if test ! -d \"\$progdir\"; then $MKDIR \"\$progdir\" else $RM \"\$progdir/\$file\" fi" $ECHO "\ # relink executable if necessary if test -n \"\$relink_command\"; then if relink_command_output=\`eval \$relink_command 2>&1\`; then : else \$ECHO \"\$relink_command_output\" >&2 $RM \"\$progdir/\$file\" exit 1 fi fi $MV \"\$progdir/\$file\" \"\$progdir/\$program\" 2>/dev/null || { $RM \"\$progdir/\$program\"; $MV \"\$progdir/\$file\" \"\$progdir/\$program\"; } $RM \"\$progdir/\$file\" fi" else $ECHO "\ program='$outputname' progdir=\"\$thisdir/$objdir\" " fi $ECHO "\ if test -f \"\$progdir/\$program\"; then" # fixup the dll searchpath if we need to. # # Fix the DLL searchpath if we need to. Do this before prepending # to shlibpath, because on Windows, both are PATH and uninstalled # libraries must come first. if test -n "$dllsearchpath"; then $ECHO "\ # Add the dll search path components to the executable PATH PATH=$dllsearchpath:\$PATH " fi # Export our shlibpath_var if we have one. if test yes = "$shlibpath_overrides_runpath" && test -n "$shlibpath_var" && test -n "$temp_rpath"; then $ECHO "\ # Add our own library path to $shlibpath_var $shlibpath_var=\"$temp_rpath\$$shlibpath_var\" # Some systems cannot cope with colon-terminated $shlibpath_var # The second colon is a workaround for a bug in BeOS R4 sed $shlibpath_var=\`\$ECHO \"\$$shlibpath_var\" | $SED 's/::*\$//'\` export $shlibpath_var " fi $ECHO "\ if test \"\$libtool_execute_magic\" != \"$magic\"; then # Run the actual program with our arguments. func_exec_program \${1+\"\$@\"} fi else # The program doesn't exist. \$ECHO \"\$0: error: '\$progdir/\$program' does not exist\" 1>&2 \$ECHO \"This script is just a wrapper for \$program.\" 1>&2 \$ECHO \"See the $PACKAGE documentation for more information.\" 1>&2 exit 1 fi fi\ " } # func_emit_cwrapperexe_src # emit the source code for a wrapper executable on stdout # Must ONLY be called from within func_mode_link because # it depends on a number of variable set therein. func_emit_cwrapperexe_src () { cat <<EOF /* $cwrappersource - temporary wrapper executable for $objdir/$outputname Generated by $PROGRAM (GNU $PACKAGE) $VERSION The $output program cannot be directly executed until all the libtool libraries that it depends on are installed. This wrapper executable should never be moved out of the build directory. If it is, it will not operate correctly. */ EOF cat <<"EOF" #ifdef _MSC_VER # define _CRT_SECURE_NO_DEPRECATE 1 #endif #include <stdio.h> #include <stdlib.h> #ifdef _MSC_VER # include <direct.h> # include <process.h> # include <io.h> #else # include <unistd.h> # include <stdint.h> # ifdef __CYGWIN__ # include <io.h> # endif #endif #include <malloc.h> #include <stdarg.h> #include <assert.h> #include <string.h> #include <ctype.h> #include <errno.h> #include <fcntl.h> #include <sys/stat.h> #define STREQ(s1, s2) (strcmp ((s1), (s2)) == 0) /* declarations of non-ANSI functions */ #if defined __MINGW32__ # ifdef __STRICT_ANSI__ int _putenv (const char *); # endif #elif defined __CYGWIN__ # ifdef __STRICT_ANSI__ char *realpath (const char *, char *); int putenv (char *); int setenv (const char *, const char *, int); # endif /* #elif defined other_platform || defined ... */ #endif /* portability defines, excluding path handling macros */ #if defined _MSC_VER # define setmode _setmode # define stat _stat # define chmod _chmod # define getcwd _getcwd # define putenv _putenv # define S_IXUSR _S_IEXEC #elif defined __MINGW32__ # define setmode _setmode # define stat _stat # define chmod _chmod # define getcwd _getcwd # define putenv _putenv #elif defined __CYGWIN__ # define HAVE_SETENV # define FOPEN_WB "wb" /* #elif defined other platforms ... */ #endif #if defined PATH_MAX # define LT_PATHMAX PATH_MAX #elif defined MAXPATHLEN # define LT_PATHMAX MAXPATHLEN #else # define LT_PATHMAX 1024 #endif #ifndef S_IXOTH # define S_IXOTH 0 #endif #ifndef S_IXGRP # define S_IXGRP 0 #endif /* path handling portability macros */ #ifndef DIR_SEPARATOR # define DIR_SEPARATOR '/' # define PATH_SEPARATOR ':' #endif #if defined _WIN32 || defined __MSDOS__ || defined __DJGPP__ || \ defined __OS2__ # define HAVE_DOS_BASED_FILE_SYSTEM # define FOPEN_WB "wb" # ifndef DIR_SEPARATOR_2 # define DIR_SEPARATOR_2 '\\' # endif # ifndef PATH_SEPARATOR_2 # define PATH_SEPARATOR_2 ';' # endif #endif #ifndef DIR_SEPARATOR_2 # define IS_DIR_SEPARATOR(ch) ((ch) == DIR_SEPARATOR) #else /* DIR_SEPARATOR_2 */ # define IS_DIR_SEPARATOR(ch) \ (((ch) == DIR_SEPARATOR) || ((ch) == DIR_SEPARATOR_2)) #endif /* DIR_SEPARATOR_2 */ #ifndef PATH_SEPARATOR_2 # define IS_PATH_SEPARATOR(ch) ((ch) == PATH_SEPARATOR) #else /* PATH_SEPARATOR_2 */ # define IS_PATH_SEPARATOR(ch) ((ch) == PATH_SEPARATOR_2) #endif /* PATH_SEPARATOR_2 */ #ifndef FOPEN_WB # define FOPEN_WB "w" #endif #ifndef _O_BINARY # define _O_BINARY 0 #endif #define XMALLOC(type, num) ((type *) xmalloc ((num) * sizeof(type))) #define XFREE(stale) do { \ if (stale) { free (stale); stale = 0; } \ } while (0) #if defined LT_DEBUGWRAPPER static int lt_debug = 1; #else static int lt_debug = 0; #endif const char *program_name = "libtool-wrapper"; /* in case xstrdup fails */ void *xmalloc (size_t num); char *xstrdup (const char *string); const char *base_name (const char *name); char *find_executable (const char *wrapper); char *chase_symlinks (const char *pathspec); int make_executable (const char *path); int check_executable (const char *path); char *strendzap (char *str, const char *pat); void lt_debugprintf (const char *file, int line, const char *fmt, ...); void lt_fatal (const char *file, int line, const char *message, ...); static const char *nonnull (const char *s); static const char *nonempty (const char *s); void lt_setenv (const char *name, const char *value); char *lt_extend_str (const char *orig_value, const char *add, int to_end); void lt_update_exe_path (const char *name, const char *value); void lt_update_lib_path (const char *name, const char *value); char **prepare_spawn (char **argv); void lt_dump_script (FILE *f); EOF cat <<EOF #if __GNUC__ < 4 || (__GNUC__ == 4 && __GNUC_MINOR__ < 5) # define externally_visible volatile #else # define externally_visible __attribute__((externally_visible)) volatile #endif externally_visible const char * MAGIC_EXE = "$magic_exe"; const char * LIB_PATH_VARNAME = "$shlibpath_var"; EOF if test yes = "$shlibpath_overrides_runpath" && test -n "$shlibpath_var" && test -n "$temp_rpath"; then func_to_host_path "$temp_rpath" cat <<EOF const char * LIB_PATH_VALUE = "$func_to_host_path_result"; EOF else cat <<"EOF" const char * LIB_PATH_VALUE = ""; EOF fi if test -n "$dllsearchpath"; then func_to_host_path "$dllsearchpath:" cat <<EOF const char * EXE_PATH_VARNAME = "PATH"; const char * EXE_PATH_VALUE = "$func_to_host_path_result"; EOF else cat <<"EOF" const char * EXE_PATH_VARNAME = ""; const char * EXE_PATH_VALUE = ""; EOF fi if test yes = "$fast_install"; then cat <<EOF const char * TARGET_PROGRAM_NAME = "lt-$outputname"; /* hopefully, no .exe */ EOF else cat <<EOF const char * TARGET_PROGRAM_NAME = "$outputname"; /* hopefully, no .exe */ EOF fi cat <<"EOF" #define LTWRAPPER_OPTION_PREFIX "--lt-" static const char *ltwrapper_option_prefix = LTWRAPPER_OPTION_PREFIX; static const char *dumpscript_opt = LTWRAPPER_OPTION_PREFIX "dump-script"; static const char *debug_opt = LTWRAPPER_OPTION_PREFIX "debug"; int main (int argc, char *argv[]) { char **newargz; int newargc; char *tmp_pathspec; char *actual_cwrapper_path; char *actual_cwrapper_name; char *target_name; char *lt_argv_zero; int rval = 127; int i; program_name = (char *) xstrdup (base_name (argv[0])); newargz = XMALLOC (char *, (size_t) argc + 1); /* very simple arg parsing; don't want to rely on getopt * also, copy all non cwrapper options to newargz, except * argz[0], which is handled differently */ newargc=0; for (i = 1; i < argc; i++) { if (STREQ (argv[i], dumpscript_opt)) { EOF case $host in *mingw* | *cygwin* ) # make stdout use "unix" line endings echo " setmode(1,_O_BINARY);" ;; esac cat <<"EOF" lt_dump_script (stdout); return 0; } if (STREQ (argv[i], debug_opt)) { lt_debug = 1; continue; } if (STREQ (argv[i], ltwrapper_option_prefix)) { /* however, if there is an option in the LTWRAPPER_OPTION_PREFIX namespace, but it is not one of the ones we know about and have already dealt with, above (inluding dump-script), then report an error. Otherwise, targets might begin to believe they are allowed to use options in the LTWRAPPER_OPTION_PREFIX namespace. The first time any user complains about this, we'll need to make LTWRAPPER_OPTION_PREFIX a configure-time option or a configure.ac-settable value. */ lt_fatal (__FILE__, __LINE__, "unrecognized %s option: '%s'", ltwrapper_option_prefix, argv[i]); } /* otherwise ... */ newargz[++newargc] = xstrdup (argv[i]); } newargz[++newargc] = NULL; EOF cat <<EOF /* The GNU banner must be the first non-error debug message */ lt_debugprintf (__FILE__, __LINE__, "libtool wrapper (GNU $PACKAGE) $VERSION\n"); EOF cat <<"EOF" lt_debugprintf (__FILE__, __LINE__, "(main) argv[0]: %s\n", argv[0]); lt_debugprintf (__FILE__, __LINE__, "(main) program_name: %s\n", program_name); tmp_pathspec = find_executable (argv[0]); if (tmp_pathspec == NULL) lt_fatal (__FILE__, __LINE__, "couldn't find %s", argv[0]); lt_debugprintf (__FILE__, __LINE__, "(main) found exe (before symlink chase) at: %s\n", tmp_pathspec); actual_cwrapper_path = chase_symlinks (tmp_pathspec); lt_debugprintf (__FILE__, __LINE__, "(main) found exe (after symlink chase) at: %s\n", actual_cwrapper_path); XFREE (tmp_pathspec); actual_cwrapper_name = xstrdup (base_name (actual_cwrapper_path)); strendzap (actual_cwrapper_path, actual_cwrapper_name); /* wrapper name transforms */ strendzap (actual_cwrapper_name, ".exe"); tmp_pathspec = lt_extend_str (actual_cwrapper_name, ".exe", 1); XFREE (actual_cwrapper_name); actual_cwrapper_name = tmp_pathspec; tmp_pathspec = 0; /* target_name transforms -- use actual target program name; might have lt- prefix */ target_name = xstrdup (base_name (TARGET_PROGRAM_NAME)); strendzap (target_name, ".exe"); tmp_pathspec = lt_extend_str (target_name, ".exe", 1); XFREE (target_name); target_name = tmp_pathspec; tmp_pathspec = 0; lt_debugprintf (__FILE__, __LINE__, "(main) libtool target name: %s\n", target_name); EOF cat <<EOF newargz[0] = XMALLOC (char, (strlen (actual_cwrapper_path) + strlen ("$objdir") + 1 + strlen (actual_cwrapper_name) + 1)); strcpy (newargz[0], actual_cwrapper_path); strcat (newargz[0], "$objdir"); strcat (newargz[0], "/"); EOF cat <<"EOF" /* stop here, and copy so we don't have to do this twice */ tmp_pathspec = xstrdup (newargz[0]); /* do NOT want the lt- prefix here, so use actual_cwrapper_name */ strcat (newargz[0], actual_cwrapper_name); /* DO want the lt- prefix here if it exists, so use target_name */ lt_argv_zero = lt_extend_str (tmp_pathspec, target_name, 1); XFREE (tmp_pathspec); tmp_pathspec = NULL; EOF case $host_os in mingw*) cat <<"EOF" { char* p; while ((p = strchr (newargz[0], '\\')) != NULL) { *p = '/'; } while ((p = strchr (lt_argv_zero, '\\')) != NULL) { *p = '/'; } } EOF ;; esac cat <<"EOF" XFREE (target_name); XFREE (actual_cwrapper_path); XFREE (actual_cwrapper_name); lt_setenv ("BIN_SH", "xpg4"); /* for Tru64 */ lt_setenv ("DUALCASE", "1"); /* for MSK sh */ /* Update the DLL searchpath. EXE_PATH_VALUE ($dllsearchpath) must be prepended before (that is, appear after) LIB_PATH_VALUE ($temp_rpath) because on Windows, both *_VARNAMEs are PATH but uninstalled libraries must come first. */ lt_update_exe_path (EXE_PATH_VARNAME, EXE_PATH_VALUE); lt_update_lib_path (LIB_PATH_VARNAME, LIB_PATH_VALUE); lt_debugprintf (__FILE__, __LINE__, "(main) lt_argv_zero: %s\n", nonnull (lt_argv_zero)); for (i = 0; i < newargc; i++) { lt_debugprintf (__FILE__, __LINE__, "(main) newargz[%d]: %s\n", i, nonnull (newargz[i])); } EOF case $host_os in mingw*) cat <<"EOF" /* execv doesn't actually work on mingw as expected on unix */ newargz = prepare_spawn (newargz); rval = (int) _spawnv (_P_WAIT, lt_argv_zero, (const char * const *) newargz); if (rval == -1) { /* failed to start process */ lt_debugprintf (__FILE__, __LINE__, "(main) failed to launch target \"%s\": %s\n", lt_argv_zero, nonnull (strerror (errno))); return 127; } return rval; EOF ;; *) cat <<"EOF" execv (lt_argv_zero, newargz); return rval; /* =127, but avoids unused variable warning */ EOF ;; esac cat <<"EOF" } void * xmalloc (size_t num) { void *p = (void *) malloc (num); if (!p) lt_fatal (__FILE__, __LINE__, "memory exhausted"); return p; } char * xstrdup (const char *string) { return string ? strcpy ((char *) xmalloc (strlen (string) + 1), string) : NULL; } const char * base_name (const char *name) { const char *base; #if defined HAVE_DOS_BASED_FILE_SYSTEM /* Skip over the disk name in MSDOS pathnames. */ if (isalpha ((unsigned char) name[0]) && name[1] == ':') name += 2; #endif for (base = name; *name; name++) if (IS_DIR_SEPARATOR (*name)) base = name + 1; return base; } int check_executable (const char *path) { struct stat st; lt_debugprintf (__FILE__, __LINE__, "(check_executable): %s\n", nonempty (path)); if ((!path) || (!*path)) return 0; if ((stat (path, &st) >= 0) && (st.st_mode & (S_IXUSR | S_IXGRP | S_IXOTH))) return 1; else return 0; } int make_executable (const char *path) { int rval = 0; struct stat st; lt_debugprintf (__FILE__, __LINE__, "(make_executable): %s\n", nonempty (path)); if ((!path) || (!*path)) return 0; if (stat (path, &st) >= 0) { rval = chmod (path, st.st_mode | S_IXOTH | S_IXGRP | S_IXUSR); } return rval; } /* Searches for the full path of the wrapper. Returns newly allocated full path name if found, NULL otherwise Does not chase symlinks, even on platforms that support them. */ char * find_executable (const char *wrapper) { int has_slash = 0; const char *p; const char *p_next; /* static buffer for getcwd */ char tmp[LT_PATHMAX + 1]; size_t tmp_len; char *concat_name; lt_debugprintf (__FILE__, __LINE__, "(find_executable): %s\n", nonempty (wrapper)); if ((wrapper == NULL) || (*wrapper == '\0')) return NULL; /* Absolute path? */ #if defined HAVE_DOS_BASED_FILE_SYSTEM if (isalpha ((unsigned char) wrapper[0]) && wrapper[1] == ':') { concat_name = xstrdup (wrapper); if (check_executable (concat_name)) return concat_name; XFREE (concat_name); } else { #endif if (IS_DIR_SEPARATOR (wrapper[0])) { concat_name = xstrdup (wrapper); if (check_executable (concat_name)) return concat_name; XFREE (concat_name); } #if defined HAVE_DOS_BASED_FILE_SYSTEM } #endif for (p = wrapper; *p; p++) if (*p == '/') { has_slash = 1; break; } if (!has_slash) { /* no slashes; search PATH */ const char *path = getenv ("PATH"); if (path != NULL) { for (p = path; *p; p = p_next) { const char *q; size_t p_len; for (q = p; *q; q++) if (IS_PATH_SEPARATOR (*q)) break; p_len = (size_t) (q - p); p_next = (*q == '\0' ? q : q + 1); if (p_len == 0) { /* empty path: current directory */ if (getcwd (tmp, LT_PATHMAX) == NULL) lt_fatal (__FILE__, __LINE__, "getcwd failed: %s", nonnull (strerror (errno))); tmp_len = strlen (tmp); concat_name = XMALLOC (char, tmp_len + 1 + strlen (wrapper) + 1); memcpy (concat_name, tmp, tmp_len); concat_name[tmp_len] = '/'; strcpy (concat_name + tmp_len + 1, wrapper); } else { concat_name = XMALLOC (char, p_len + 1 + strlen (wrapper) + 1); memcpy (concat_name, p, p_len); concat_name[p_len] = '/'; strcpy (concat_name + p_len + 1, wrapper); } if (check_executable (concat_name)) return concat_name; XFREE (concat_name); } } /* not found in PATH; assume curdir */ } /* Relative path | not found in path: prepend cwd */ if (getcwd (tmp, LT_PATHMAX) == NULL) lt_fatal (__FILE__, __LINE__, "getcwd failed: %s", nonnull (strerror (errno))); tmp_len = strlen (tmp); concat_name = XMALLOC (char, tmp_len + 1 + strlen (wrapper) + 1); memcpy (concat_name, tmp, tmp_len); concat_name[tmp_len] = '/'; strcpy (concat_name + tmp_len + 1, wrapper); if (check_executable (concat_name)) return concat_name; XFREE (concat_name); return NULL; } char * chase_symlinks (const char *pathspec) { #ifndef S_ISLNK return xstrdup (pathspec); #else char buf[LT_PATHMAX]; struct stat s; char *tmp_pathspec = xstrdup (pathspec); char *p; int has_symlinks = 0; while (strlen (tmp_pathspec) && !has_symlinks) { lt_debugprintf (__FILE__, __LINE__, "checking path component for symlinks: %s\n", tmp_pathspec); if (lstat (tmp_pathspec, &s) == 0) { if (S_ISLNK (s.st_mode) != 0) { has_symlinks = 1; break; } /* search backwards for last DIR_SEPARATOR */ p = tmp_pathspec + strlen (tmp_pathspec) - 1; while ((p > tmp_pathspec) && (!IS_DIR_SEPARATOR (*p))) p--; if ((p == tmp_pathspec) && (!IS_DIR_SEPARATOR (*p))) { /* no more DIR_SEPARATORS left */ break; } *p = '\0'; } else { lt_fatal (__FILE__, __LINE__, "error accessing file \"%s\": %s", tmp_pathspec, nonnull (strerror (errno))); } } XFREE (tmp_pathspec); if (!has_symlinks) { return xstrdup (pathspec); } tmp_pathspec = realpath (pathspec, buf); if (tmp_pathspec == 0) { lt_fatal (__FILE__, __LINE__, "could not follow symlinks for %s", pathspec); } return xstrdup (tmp_pathspec); #endif } char * strendzap (char *str, const char *pat) { size_t len, patlen; assert (str != NULL); assert (pat != NULL); len = strlen (str); patlen = strlen (pat); if (patlen <= len) { str += len - patlen; if (STREQ (str, pat)) *str = '\0'; } return str; } void lt_debugprintf (const char *file, int line, const char *fmt, ...) { va_list args; if (lt_debug) { (void) fprintf (stderr, "%s:%s:%d: ", program_name, file, line); va_start (args, fmt); (void) vfprintf (stderr, fmt, args); va_end (args); } } static void lt_error_core (int exit_status, const char *file, int line, const char *mode, const char *message, va_list ap) { fprintf (stderr, "%s:%s:%d: %s: ", program_name, file, line, mode); vfprintf (stderr, message, ap); fprintf (stderr, ".\n"); if (exit_status >= 0) exit (exit_status); } void lt_fatal (const char *file, int line, const char *message, ...) { va_list ap; va_start (ap, message); lt_error_core (EXIT_FAILURE, file, line, "FATAL", message, ap); va_end (ap); } static const char * nonnull (const char *s) { return s ? s : "(null)"; } static const char * nonempty (const char *s) { return (s && !*s) ? "(empty)" : nonnull (s); } void lt_setenv (const char *name, const char *value) { lt_debugprintf (__FILE__, __LINE__, "(lt_setenv) setting '%s' to '%s'\n", nonnull (name), nonnull (value)); { #ifdef HAVE_SETENV /* always make a copy, for consistency with !HAVE_SETENV */ char *str = xstrdup (value); setenv (name, str, 1); #else size_t len = strlen (name) + 1 + strlen (value) + 1; char *str = XMALLOC (char, len); sprintf (str, "%s=%s", name, value); if (putenv (str) != EXIT_SUCCESS) { XFREE (str); } #endif } } char * lt_extend_str (const char *orig_value, const char *add, int to_end) { char *new_value; if (orig_value && *orig_value) { size_t orig_value_len = strlen (orig_value); size_t add_len = strlen (add); new_value = XMALLOC (char, add_len + orig_value_len + 1); if (to_end) { strcpy (new_value, orig_value); strcpy (new_value + orig_value_len, add); } else { strcpy (new_value, add); strcpy (new_value + add_len, orig_value); } } else { new_value = xstrdup (add); } return new_value; } void lt_update_exe_path (const char *name, const char *value) { lt_debugprintf (__FILE__, __LINE__, "(lt_update_exe_path) modifying '%s' by prepending '%s'\n", nonnull (name), nonnull (value)); if (name && *name && value && *value) { char *new_value = lt_extend_str (getenv (name), value, 0); /* some systems can't cope with a ':'-terminated path #' */ size_t len = strlen (new_value); while ((len > 0) && IS_PATH_SEPARATOR (new_value[len-1])) { new_value[--len] = '\0'; } lt_setenv (name, new_value); XFREE (new_value); } } void lt_update_lib_path (const char *name, const char *value) { lt_debugprintf (__FILE__, __LINE__, "(lt_update_lib_path) modifying '%s' by prepending '%s'\n", nonnull (name), nonnull (value)); if (name && *name && value && *value) { char *new_value = lt_extend_str (getenv (name), value, 0); lt_setenv (name, new_value); XFREE (new_value); } } EOF case $host_os in mingw*) cat <<"EOF" /* Prepares an argument vector before calling spawn(). Note that spawn() does not by itself call the command interpreter (getenv ("COMSPEC") != NULL ? getenv ("COMSPEC") : ({ OSVERSIONINFO v; v.dwOSVersionInfoSize = sizeof(OSVERSIONINFO); GetVersionEx(&v); v.dwPlatformId == VER_PLATFORM_WIN32_NT; }) ? "cmd.exe" : "command.com"). Instead it simply concatenates the arguments, separated by ' ', and calls CreateProcess(). We must quote the arguments since Win32 CreateProcess() interprets characters like ' ', '\t', '\\', '"' (but not '<' and '>') in a special way: - Space and tab are interpreted as delimiters. They are not treated as delimiters if they are surrounded by double quotes: "...". - Unescaped double quotes are removed from the input. Their only effect is that within double quotes, space and tab are treated like normal characters. - Backslashes not followed by double quotes are not special. - But 2*n+1 backslashes followed by a double quote become n backslashes followed by a double quote (n >= 0): \" -> " \\\" -> \" \\\\\" -> \\" */ #define SHELL_SPECIAL_CHARS "\"\\ \001\002\003\004\005\006\007\010\011\012\013\014\015\016\017\020\021\022\023\024\025\026\027\030\031\032\033\034\035\036\037" #define SHELL_SPACE_CHARS " \001\002\003\004\005\006\007\010\011\012\013\014\015\016\017\020\021\022\023\024\025\026\027\030\031\032\033\034\035\036\037" char ** prepare_spawn (char **argv) { size_t argc; char **new_argv; size_t i; /* Count number of arguments. */ for (argc = 0; argv[argc] != NULL; argc++) ; /* Allocate new argument vector. */ new_argv = XMALLOC (char *, argc + 1); /* Put quoted arguments into the new argument vector. */ for (i = 0; i < argc; i++) { const char *string = argv[i]; if (string[0] == '\0') new_argv[i] = xstrdup ("\"\""); else if (strpbrk (string, SHELL_SPECIAL_CHARS) != NULL) { int quote_around = (strpbrk (string, SHELL_SPACE_CHARS) != NULL); size_t length; unsigned int backslashes; const char *s; char *quoted_string; char *p; length = 0; backslashes = 0; if (quote_around) length++; for (s = string; *s != '\0'; s++) { char c = *s; if (c == '"') length += backslashes + 1; length++; if (c == '\\') backslashes++; else backslashes = 0; } if (quote_around) length += backslashes + 1; quoted_string = XMALLOC (char, length + 1); p = quoted_string; backslashes = 0; if (quote_around) *p++ = '"'; for (s = string; *s != '\0'; s++) { char c = *s; if (c == '"') { unsigned int j; for (j = backslashes + 1; j > 0; j--) *p++ = '\\'; } *p++ = c; if (c == '\\') backslashes++; else backslashes = 0; } if (quote_around) { unsigned int j; for (j = backslashes; j > 0; j--) *p++ = '\\'; *p++ = '"'; } *p = '\0'; new_argv[i] = quoted_string; } else new_argv[i] = (char *) string; } new_argv[argc] = NULL; return new_argv; } EOF ;; esac cat <<"EOF" void lt_dump_script (FILE* f) { EOF func_emit_wrapper yes | $SED -n -e ' s/^\(.\{79\}\)\(..*\)/\1\ \2/ h s/\([\\"]\)/\\\1/g s/$/\\n/ s/\([^\n]*\).*/ fputs ("\1", f);/p g D' cat <<"EOF" } EOF } # end: func_emit_cwrapperexe_src # func_win32_import_lib_p ARG # True if ARG is an import lib, as indicated by $file_magic_cmd func_win32_import_lib_p () { $debug_cmd case `eval $file_magic_cmd \"\$1\" 2>/dev/null | $SED -e 10q` in *import*) : ;; *) false ;; esac } # func_suncc_cstd_abi # !!ONLY CALL THIS FOR SUN CC AFTER $compile_command IS FULLY EXPANDED!! # Several compiler flags select an ABI that is incompatible with the # Cstd library. Avoid specifying it if any are in CXXFLAGS. func_suncc_cstd_abi () { $debug_cmd case " $compile_command " in *" -compat=g "*|*\ -std=c++[0-9][0-9]\ *|*" -library=stdcxx4 "*|*" -library=stlport4 "*) suncc_use_cstd_abi=no ;; *) suncc_use_cstd_abi=yes ;; esac } # func_mode_link arg... func_mode_link () { $debug_cmd case $host in *-*-cygwin* | *-*-mingw* | *-*-pw32* | *-*-os2* | *-cegcc*) # It is impossible to link a dll without this setting, and # we shouldn't force the makefile maintainer to figure out # what system we are compiling for in order to pass an extra # flag for every libtool invocation. # allow_undefined=no # FIXME: Unfortunately, there are problems with the above when trying # to make a dll that has undefined symbols, in which case not # even a static library is built. For now, we need to specify # -no-undefined on the libtool link line when we can be certain # that all symbols are satisfied, otherwise we get a static library. allow_undefined=yes ;; *) allow_undefined=yes ;; esac libtool_args=$nonopt base_compile="$nonopt $@" compile_command=$nonopt finalize_command=$nonopt compile_rpath= finalize_rpath= compile_shlibpath= finalize_shlibpath= convenience= old_convenience= deplibs= old_deplibs= compiler_flags= linker_flags= dllsearchpath= lib_search_path=`pwd` inst_prefix_dir= new_inherited_linker_flags= avoid_version=no bindir= dlfiles= dlprefiles= dlself=no export_dynamic=no export_symbols= export_symbols_regex= generated= libobjs= ltlibs= module=no no_install=no objs= os2dllname= non_pic_objects= precious_files_regex= prefer_static_libs=no preload=false prev= prevarg= release= rpath= xrpath= perm_rpath= temp_rpath= thread_safe=no vinfo= vinfo_number=no weak_libs= single_module=$wl-single_module func_infer_tag $base_compile # We need to know -static, to get the right output filenames. for arg do case $arg in -shared) test yes != "$build_libtool_libs" \ && func_fatal_configuration "cannot build a shared library" build_old_libs=no break ;; -all-static | -static | -static-libtool-libs) case $arg in -all-static) if test yes = "$build_libtool_libs" && test -z "$link_static_flag"; then func_warning "complete static linking is impossible in this configuration" fi if test -n "$link_static_flag"; then dlopen_self=$dlopen_self_static fi prefer_static_libs=yes ;; -static) if test -z "$pic_flag" && test -n "$link_static_flag"; then dlopen_self=$dlopen_self_static fi prefer_static_libs=built ;; -static-libtool-libs) if test -z "$pic_flag" && test -n "$link_static_flag"; then dlopen_self=$dlopen_self_static fi prefer_static_libs=yes ;; esac build_libtool_libs=no build_old_libs=yes break ;; esac done # See if our shared archives depend on static archives. test -n "$old_archive_from_new_cmds" && build_old_libs=yes # Go through the arguments, transforming them on the way. while test "$#" -gt 0; do arg=$1 shift func_quote_for_eval "$arg" qarg=$func_quote_for_eval_unquoted_result func_append libtool_args " $func_quote_for_eval_result" # If the previous option needs an argument, assign it. if test -n "$prev"; then case $prev in output) func_append compile_command " @OUTPUT@" func_append finalize_command " @OUTPUT@" ;; esac case $prev in bindir) bindir=$arg prev= continue ;; dlfiles|dlprefiles) $preload || { # Add the symbol object into the linking commands. func_append compile_command " @SYMFILE@" func_append finalize_command " @SYMFILE@" preload=: } case $arg in *.la | *.lo) ;; # We handle these cases below. force) if test no = "$dlself"; then dlself=needless export_dynamic=yes fi prev= continue ;; self) if test dlprefiles = "$prev"; then dlself=yes elif test dlfiles = "$prev" && test yes != "$dlopen_self"; then dlself=yes else dlself=needless export_dynamic=yes fi prev= continue ;; *) if test dlfiles = "$prev"; then func_append dlfiles " $arg" else func_append dlprefiles " $arg" fi prev= continue ;; esac ;; expsyms) export_symbols=$arg test -f "$arg" \ || func_fatal_error "symbol file '$arg' does not exist" prev= continue ;; expsyms_regex) export_symbols_regex=$arg prev= continue ;; framework) case $host in *-*-darwin*) case "$deplibs " in *" $qarg.ltframework "*) ;; *) func_append deplibs " $qarg.ltframework" # this is fixed later ;; esac ;; esac prev= continue ;; inst_prefix) inst_prefix_dir=$arg prev= continue ;; mllvm) # Clang does not use LLVM to link, so we can simply discard any # '-mllvm $arg' options when doing the link step. prev= continue ;; objectlist) if test -f "$arg"; then save_arg=$arg moreargs= for fil in `cat "$save_arg"` do # func_append moreargs " $fil" arg=$fil # A libtool-controlled object. # Check to see that this really is a libtool object. if func_lalib_unsafe_p "$arg"; then pic_object= non_pic_object= # Read the .lo file func_source "$arg" if test -z "$pic_object" || test -z "$non_pic_object" || test none = "$pic_object" && test none = "$non_pic_object"; then func_fatal_error "cannot find name of object for '$arg'" fi # Extract subdirectory from the argument. func_dirname "$arg" "/" "" xdir=$func_dirname_result if test none != "$pic_object"; then # Prepend the subdirectory the object is found in. pic_object=$xdir$pic_object if test dlfiles = "$prev"; then if test yes = "$build_libtool_libs" && test yes = "$dlopen_support"; then func_append dlfiles " $pic_object" prev= continue else # If libtool objects are unsupported, then we need to preload. prev=dlprefiles fi fi # CHECK ME: I think I busted this. -Ossama if test dlprefiles = "$prev"; then # Preload the old-style object. func_append dlprefiles " $pic_object" prev= fi # A PIC object. func_append libobjs " $pic_object" arg=$pic_object fi # Non-PIC object. if test none != "$non_pic_object"; then # Prepend the subdirectory the object is found in. non_pic_object=$xdir$non_pic_object # A standard non-PIC object func_append non_pic_objects " $non_pic_object" if test -z "$pic_object" || test none = "$pic_object"; then arg=$non_pic_object fi else # If the PIC object exists, use it instead. # $xdir was prepended to $pic_object above. non_pic_object=$pic_object func_append non_pic_objects " $non_pic_object" fi else # Only an error if not doing a dry-run. if $opt_dry_run; then # Extract subdirectory from the argument. func_dirname "$arg" "/" "" xdir=$func_dirname_result func_lo2o "$arg" pic_object=$xdir$objdir/$func_lo2o_result non_pic_object=$xdir$func_lo2o_result func_append libobjs " $pic_object" func_append non_pic_objects " $non_pic_object" else func_fatal_error "'$arg' is not a valid libtool object" fi fi done else func_fatal_error "link input file '$arg' does not exist" fi arg=$save_arg prev= continue ;; os2dllname) os2dllname=$arg prev= continue ;; precious_regex) precious_files_regex=$arg prev= continue ;; release) release=-$arg prev= continue ;; rpath | xrpath) # We need an absolute path. case $arg in [\\/]* | [A-Za-z]:[\\/]*) ;; *) func_fatal_error "only absolute run-paths are allowed" ;; esac if test rpath = "$prev"; then case "$rpath " in *" $arg "*) ;; *) func_append rpath " $arg" ;; esac else case "$xrpath " in *" $arg "*) ;; *) func_append xrpath " $arg" ;; esac fi prev= continue ;; shrext) shrext_cmds=$arg prev= continue ;; weak) func_append weak_libs " $arg" prev= continue ;; xcclinker) func_append linker_flags " $qarg" func_append compiler_flags " $qarg" prev= func_append compile_command " $qarg" func_append finalize_command " $qarg" continue ;; xcompiler) func_append compiler_flags " $qarg" prev= func_append compile_command " $qarg" func_append finalize_command " $qarg" continue ;; xlinker) func_append linker_flags " $qarg" func_append compiler_flags " $wl$qarg" prev= func_append compile_command " $wl$qarg" func_append finalize_command " $wl$qarg" continue ;; *) eval "$prev=\"\$arg\"" prev= continue ;; esac fi # test -n "$prev" prevarg=$arg case $arg in -all-static) if test -n "$link_static_flag"; then # See comment for -static flag below, for more details. func_append compile_command " $link_static_flag" func_append finalize_command " $link_static_flag" fi continue ;; -allow-undefined) # FIXME: remove this flag sometime in the future. func_fatal_error "'-allow-undefined' must not be used because it is the default" ;; -avoid-version) avoid_version=yes continue ;; -bindir) prev=bindir continue ;; -dlopen) prev=dlfiles continue ;; -dlpreopen) prev=dlprefiles continue ;; -export-dynamic) export_dynamic=yes continue ;; -export-symbols | -export-symbols-regex) if test -n "$export_symbols" || test -n "$export_symbols_regex"; then func_fatal_error "more than one -exported-symbols argument is not allowed" fi if test X-export-symbols = "X$arg"; then prev=expsyms else prev=expsyms_regex fi continue ;; -framework) prev=framework continue ;; -inst-prefix-dir) prev=inst_prefix continue ;; # The native IRIX linker understands -LANG:*, -LIST:* and -LNO:* # so, if we see these flags be careful not to treat them like -L -L[A-Z][A-Z]*:*) case $with_gcc/$host in no/*-*-irix* | /*-*-irix*) func_append compile_command " $arg" func_append finalize_command " $arg" ;; esac continue ;; -L*) func_stripname "-L" '' "$arg" if test -z "$func_stripname_result"; then if test "$#" -gt 0; then func_fatal_error "require no space between '-L' and '$1'" else func_fatal_error "need path for '-L' option" fi fi func_resolve_sysroot "$func_stripname_result" dir=$func_resolve_sysroot_result # We need an absolute path. case $dir in [\\/]* | [A-Za-z]:[\\/]*) ;; *) absdir=`cd "$dir" && pwd` test -z "$absdir" && \ func_fatal_error "cannot determine absolute directory name of '$dir'" dir=$absdir ;; esac case "$deplibs " in *" -L$dir "* | *" $arg "*) # Will only happen for absolute or sysroot arguments ;; *) # Preserve sysroot, but never include relative directories case $dir in [\\/]* | [A-Za-z]:[\\/]* | =*) func_append deplibs " $arg" ;; *) func_append deplibs " -L$dir" ;; esac func_append lib_search_path " $dir" ;; esac case $host in *-*-cygwin* | *-*-mingw* | *-*-pw32* | *-*-os2* | *-cegcc*) testbindir=`$ECHO "$dir" | $SED 's*/lib$*/bin*'` case :$dllsearchpath: in *":$dir:"*) ;; ::) dllsearchpath=$dir;; *) func_append dllsearchpath ":$dir";; esac case :$dllsearchpath: in *":$testbindir:"*) ;; ::) dllsearchpath=$testbindir;; *) func_append dllsearchpath ":$testbindir";; esac ;; esac continue ;; -l*) if test X-lc = "X$arg" || test X-lm = "X$arg"; then case $host in *-*-cygwin* | *-*-mingw* | *-*-pw32* | *-*-beos* | *-cegcc* | *-*-haiku*) # These systems don't actually have a C or math library (as such) continue ;; *-*-os2*) # These systems don't actually have a C library (as such) test X-lc = "X$arg" && continue ;; *-*-openbsd* | *-*-freebsd* | *-*-dragonfly* | *-*-bitrig*) # Do not include libc due to us having libc/libc_r. test X-lc = "X$arg" && continue ;; *-*-rhapsody* | *-*-darwin1.[012]) # Rhapsody C and math libraries are in the System framework func_append deplibs " System.ltframework" continue ;; *-*-sco3.2v5* | *-*-sco5v6*) # Causes problems with __ctype test X-lc = "X$arg" && continue ;; *-*-sysv4.2uw2* | *-*-sysv5* | *-*-unixware* | *-*-OpenUNIX*) # Compiler inserts libc in the correct place for threads to work test X-lc = "X$arg" && continue ;; esac elif test X-lc_r = "X$arg"; then case $host in *-*-openbsd* | *-*-freebsd* | *-*-dragonfly* | *-*-bitrig*) # Do not include libc_r directly, use -pthread flag. continue ;; esac fi func_append deplibs " $arg" continue ;; -mllvm) prev=mllvm continue ;; -module) module=yes continue ;; # Tru64 UNIX uses -model [arg] to determine the layout of C++ # classes, name mangling, and exception handling. # Darwin uses the -arch flag to determine output architecture. -model|-arch|-isysroot|--sysroot) func_append compiler_flags " $arg" func_append compile_command " $arg" func_append finalize_command " $arg" prev=xcompiler continue ;; -mt|-mthreads|-kthread|-Kthread|-pthread|-pthreads|--thread-safe \ |-threads|-fopenmp|-openmp|-mp|-xopenmp|-omp|-qsmp=*) func_append compiler_flags " $arg" func_append compile_command " $arg" func_append finalize_command " $arg" case "$new_inherited_linker_flags " in *" $arg "*) ;; * ) func_append new_inherited_linker_flags " $arg" ;; esac continue ;; -multi_module) single_module=$wl-multi_module continue ;; -no-fast-install) fast_install=no continue ;; -no-install) case $host in *-*-cygwin* | *-*-mingw* | *-*-pw32* | *-*-os2* | *-*-darwin* | *-cegcc*) # The PATH hackery in wrapper scripts is required on Windows # and Darwin in order for the loader to find any dlls it needs. func_warning "'-no-install' is ignored for $host" func_warning "assuming '-no-fast-install' instead" fast_install=no ;; *) no_install=yes ;; esac continue ;; -no-undefined) allow_undefined=no continue ;; -objectlist) prev=objectlist continue ;; -os2dllname) prev=os2dllname continue ;; -o) prev=output ;; -precious-files-regex) prev=precious_regex continue ;; -release) prev=release continue ;; -rpath) prev=rpath continue ;; -R) prev=xrpath continue ;; -R*) func_stripname '-R' '' "$arg" dir=$func_stripname_result # We need an absolute path. case $dir in [\\/]* | [A-Za-z]:[\\/]*) ;; =*) func_stripname '=' '' "$dir" dir=$lt_sysroot$func_stripname_result ;; *) func_fatal_error "only absolute run-paths are allowed" ;; esac case "$xrpath " in *" $dir "*) ;; *) func_append xrpath " $dir" ;; esac continue ;; -shared) # The effects of -shared are defined in a previous loop. continue ;; -shrext) prev=shrext continue ;; -static | -static-libtool-libs) # The effects of -static are defined in a previous loop. # We used to do the same as -all-static on platforms that # didn't have a PIC flag, but the assumption that the effects # would be equivalent was wrong. It would break on at least # Digital Unix and AIX. continue ;; -thread-safe) thread_safe=yes continue ;; -version-info) prev=vinfo continue ;; -version-number) prev=vinfo vinfo_number=yes continue ;; -weak) prev=weak continue ;; -Wc,*) func_stripname '-Wc,' '' "$arg" args=$func_stripname_result arg= save_ifs=$IFS; IFS=, for flag in $args; do IFS=$save_ifs func_quote_for_eval "$flag" func_append arg " $func_quote_for_eval_result" func_append compiler_flags " $func_quote_for_eval_result" done IFS=$save_ifs func_stripname ' ' '' "$arg" arg=$func_stripname_result ;; -Wl,*) func_stripname '-Wl,' '' "$arg" args=$func_stripname_result arg= save_ifs=$IFS; IFS=, for flag in $args; do IFS=$save_ifs func_quote_for_eval "$flag" func_append arg " $wl$func_quote_for_eval_result" func_append compiler_flags " $wl$func_quote_for_eval_result" func_append linker_flags " $func_quote_for_eval_result" done IFS=$save_ifs func_stripname ' ' '' "$arg" arg=$func_stripname_result ;; -Xcompiler) prev=xcompiler continue ;; -Xlinker) prev=xlinker continue ;; -XCClinker) prev=xcclinker continue ;; # -msg_* for osf cc -msg_*) func_quote_for_eval "$arg" arg=$func_quote_for_eval_result ;; # Flags to be passed through unchanged, with rationale: # -64, -mips[0-9] enable 64-bit mode for the SGI compiler # -r[0-9][0-9]* specify processor for the SGI compiler # -xarch=*, -xtarget=* enable 64-bit mode for the Sun compiler # +DA*, +DD* enable 64-bit mode for the HP compiler # -q* compiler args for the IBM compiler # -m*, -t[45]*, -txscale* architecture-specific flags for GCC # -F/path path to uninstalled frameworks, gcc on darwin # -p, -pg, --coverage, -fprofile-* profiling flags for GCC # -fstack-protector* stack protector flags for GCC # @file GCC response files # -tp=* Portland pgcc target processor selection # --sysroot=* for sysroot support # -O*, -g*, -flto*, -fwhopr*, -fuse-linker-plugin GCC link-time optimization # -specs=* GCC specs files # -stdlib=* select c++ std lib with clang # -fsanitize=* Clang/GCC memory and address sanitizer -64|-mips[0-9]|-r[0-9][0-9]*|-xarch=*|-xtarget=*|+DA*|+DD*|-q*|-m*| \ -t[45]*|-txscale*|-p|-pg|--coverage|-fprofile-*|-F*|@*|-tp=*|--sysroot=*| \ -O*|-g*|-flto*|-fwhopr*|-fuse-linker-plugin|-fstack-protector*|-stdlib=*| \ -specs=*|-fsanitize=*) func_quote_for_eval "$arg" arg=$func_quote_for_eval_result func_append compile_command " $arg" func_append finalize_command " $arg" func_append compiler_flags " $arg" continue ;; -Z*) if test os2 = "`expr $host : '.*\(os2\)'`"; then # OS/2 uses -Zxxx to specify OS/2-specific options compiler_flags="$compiler_flags $arg" func_append compile_command " $arg" func_append finalize_command " $arg" case $arg in -Zlinker | -Zstack) prev=xcompiler ;; esac continue else # Otherwise treat like 'Some other compiler flag' below func_quote_for_eval "$arg" arg=$func_quote_for_eval_result fi ;; # Some other compiler flag. -* | +*) func_quote_for_eval "$arg" arg=$func_quote_for_eval_result ;; *.$objext) # A standard object. func_append objs " $arg" ;; *.lo) # A libtool-controlled object. # Check to see that this really is a libtool object. if func_lalib_unsafe_p "$arg"; then pic_object= non_pic_object= # Read the .lo file func_source "$arg" if test -z "$pic_object" || test -z "$non_pic_object" || test none = "$pic_object" && test none = "$non_pic_object"; then func_fatal_error "cannot find name of object for '$arg'" fi # Extract subdirectory from the argument. func_dirname "$arg" "/" "" xdir=$func_dirname_result test none = "$pic_object" || { # Prepend the subdirectory the object is found in. pic_object=$xdir$pic_object if test dlfiles = "$prev"; then if test yes = "$build_libtool_libs" && test yes = "$dlopen_support"; then func_append dlfiles " $pic_object" prev= continue else # If libtool objects are unsupported, then we need to preload. prev=dlprefiles fi fi # CHECK ME: I think I busted this. -Ossama if test dlprefiles = "$prev"; then # Preload the old-style object. func_append dlprefiles " $pic_object" prev= fi # A PIC object. func_append libobjs " $pic_object" arg=$pic_object } # Non-PIC object. if test none != "$non_pic_object"; then # Prepend the subdirectory the object is found in. non_pic_object=$xdir$non_pic_object # A standard non-PIC object func_append non_pic_objects " $non_pic_object" if test -z "$pic_object" || test none = "$pic_object"; then arg=$non_pic_object fi else # If the PIC object exists, use it instead. # $xdir was prepended to $pic_object above. non_pic_object=$pic_object func_append non_pic_objects " $non_pic_object" fi else # Only an error if not doing a dry-run. if $opt_dry_run; then # Extract subdirectory from the argument. func_dirname "$arg" "/" "" xdir=$func_dirname_result func_lo2o "$arg" pic_object=$xdir$objdir/$func_lo2o_result non_pic_object=$xdir$func_lo2o_result func_append libobjs " $pic_object" func_append non_pic_objects " $non_pic_object" else func_fatal_error "'$arg' is not a valid libtool object" fi fi ;; *.$libext) # An archive. func_append deplibs " $arg" func_append old_deplibs " $arg" continue ;; *.la) # A libtool-controlled library. func_resolve_sysroot "$arg" if test dlfiles = "$prev"; then # This library was specified with -dlopen. func_append dlfiles " $func_resolve_sysroot_result" prev= elif test dlprefiles = "$prev"; then # The library was specified with -dlpreopen. func_append dlprefiles " $func_resolve_sysroot_result" prev= else func_append deplibs " $func_resolve_sysroot_result" fi continue ;; # Some other compiler argument. *) # Unknown arguments in both finalize_command and compile_command need # to be aesthetically quoted because they are evaled later. func_quote_for_eval "$arg" arg=$func_quote_for_eval_result ;; esac # arg # Now actually substitute the argument into the commands. if test -n "$arg"; then func_append compile_command " $arg" func_append finalize_command " $arg" fi done # argument parsing loop test -n "$prev" && \ func_fatal_help "the '$prevarg' option requires an argument" if test yes = "$export_dynamic" && test -n "$export_dynamic_flag_spec"; then eval arg=\"$export_dynamic_flag_spec\" func_append compile_command " $arg" func_append finalize_command " $arg" fi oldlibs= # calculate the name of the file, without its directory func_basename "$output" outputname=$func_basename_result libobjs_save=$libobjs if test -n "$shlibpath_var"; then # get the directories listed in $shlibpath_var eval shlib_search_path=\`\$ECHO \"\$$shlibpath_var\" \| \$SED \'s/:/ /g\'\` else shlib_search_path= fi eval sys_lib_search_path=\"$sys_lib_search_path_spec\" eval sys_lib_dlsearch_path=\"$sys_lib_dlsearch_path_spec\" # Definition is injected by LT_CONFIG during libtool generation. func_munge_path_list sys_lib_dlsearch_path "$LT_SYS_LIBRARY_PATH" func_dirname "$output" "/" "" output_objdir=$func_dirname_result$objdir func_to_tool_file "$output_objdir/" tool_output_objdir=$func_to_tool_file_result # Create the object directory. func_mkdir_p "$output_objdir" # Determine the type of output case $output in "") func_fatal_help "you must specify an output file" ;; *.$libext) linkmode=oldlib ;; *.lo | *.$objext) linkmode=obj ;; *.la) linkmode=lib ;; *) linkmode=prog ;; # Anything else should be a program. esac specialdeplibs= libs= # Find all interdependent deplibs by searching for libraries # that are linked more than once (e.g. -la -lb -la) for deplib in $deplibs; do if $opt_preserve_dup_deps; then case "$libs " in *" $deplib "*) func_append specialdeplibs " $deplib" ;; esac fi func_append libs " $deplib" done if test lib = "$linkmode"; then libs="$predeps $libs $compiler_lib_search_path $postdeps" # Compute libraries that are listed more than once in $predeps # $postdeps and mark them as special (i.e., whose duplicates are # not to be eliminated). pre_post_deps= if $opt_duplicate_compiler_generated_deps; then for pre_post_dep in $predeps $postdeps; do case "$pre_post_deps " in *" $pre_post_dep "*) func_append specialdeplibs " $pre_post_deps" ;; esac func_append pre_post_deps " $pre_post_dep" done fi pre_post_deps= fi deplibs= newdependency_libs= newlib_search_path= need_relink=no # whether we're linking any uninstalled libtool libraries notinst_deplibs= # not-installed libtool libraries notinst_path= # paths that contain not-installed libtool libraries case $linkmode in lib) passes="conv dlpreopen link" for file in $dlfiles $dlprefiles; do case $file in *.la) ;; *) func_fatal_help "libraries can '-dlopen' only libtool libraries: $file" ;; esac done ;; prog) compile_deplibs= finalize_deplibs= alldeplibs=false newdlfiles= newdlprefiles= passes="conv scan dlopen dlpreopen link" ;; *) passes="conv" ;; esac for pass in $passes; do # The preopen pass in lib mode reverses $deplibs; put it back here # so that -L comes before libs that need it for instance... if test lib,link = "$linkmode,$pass"; then ## FIXME: Find the place where the list is rebuilt in the wrong ## order, and fix it there properly tmp_deplibs= for deplib in $deplibs; do tmp_deplibs="$deplib $tmp_deplibs" done deplibs=$tmp_deplibs fi if test lib,link = "$linkmode,$pass" || test prog,scan = "$linkmode,$pass"; then libs=$deplibs deplibs= fi if test prog = "$linkmode"; then case $pass in dlopen) libs=$dlfiles ;; dlpreopen) libs=$dlprefiles ;; link) libs="$deplibs %DEPLIBS%" test "X$link_all_deplibs" != Xno && libs="$libs $dependency_libs" ;; esac fi if test lib,dlpreopen = "$linkmode,$pass"; then # Collect and forward deplibs of preopened libtool libs for lib in $dlprefiles; do # Ignore non-libtool-libs dependency_libs= func_resolve_sysroot "$lib" case $lib in *.la) func_source "$func_resolve_sysroot_result" ;; esac # Collect preopened libtool deplibs, except any this library # has declared as weak libs for deplib in $dependency_libs; do func_basename "$deplib" deplib_base=$func_basename_result case " $weak_libs " in *" $deplib_base "*) ;; *) func_append deplibs " $deplib" ;; esac done done libs=$dlprefiles fi if test dlopen = "$pass"; then # Collect dlpreopened libraries save_deplibs=$deplibs deplibs= fi for deplib in $libs; do lib= found=false case $deplib in -mt|-mthreads|-kthread|-Kthread|-pthread|-pthreads|--thread-safe \ |-threads|-fopenmp|-openmp|-mp|-xopenmp|-omp|-qsmp=*) if test prog,link = "$linkmode,$pass"; then compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" else func_append compiler_flags " $deplib" if test lib = "$linkmode"; then case "$new_inherited_linker_flags " in *" $deplib "*) ;; * ) func_append new_inherited_linker_flags " $deplib" ;; esac fi fi continue ;; -l*) if test lib != "$linkmode" && test prog != "$linkmode"; then func_warning "'-l' is ignored for archives/objects" continue fi func_stripname '-l' '' "$deplib" name=$func_stripname_result if test lib = "$linkmode"; then searchdirs="$newlib_search_path $lib_search_path $compiler_lib_search_dirs $sys_lib_search_path $shlib_search_path" else searchdirs="$newlib_search_path $lib_search_path $sys_lib_search_path $shlib_search_path" fi for searchdir in $searchdirs; do for search_ext in .la $std_shrext .so .a; do # Search the libtool library lib=$searchdir/lib$name$search_ext if test -f "$lib"; then if test .la = "$search_ext"; then found=: else found=false fi break 2 fi done done if $found; then # deplib is a libtool library # If $allow_libtool_libs_with_static_runtimes && $deplib is a stdlib, # We need to do some special things here, and not later. if test yes = "$allow_libtool_libs_with_static_runtimes"; then case " $predeps $postdeps " in *" $deplib "*) if func_lalib_p "$lib"; then library_names= old_library= func_source "$lib" for l in $old_library $library_names; do ll=$l done if test "X$ll" = "X$old_library"; then # only static version available found=false func_dirname "$lib" "" "." ladir=$func_dirname_result lib=$ladir/$old_library if test prog,link = "$linkmode,$pass"; then compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" else deplibs="$deplib $deplibs" test lib = "$linkmode" && newdependency_libs="$deplib $newdependency_libs" fi continue fi fi ;; *) ;; esac fi else # deplib doesn't seem to be a libtool library if test prog,link = "$linkmode,$pass"; then compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" else deplibs="$deplib $deplibs" test lib = "$linkmode" && newdependency_libs="$deplib $newdependency_libs" fi continue fi ;; # -l *.ltframework) if test prog,link = "$linkmode,$pass"; then compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" else deplibs="$deplib $deplibs" if test lib = "$linkmode"; then case "$new_inherited_linker_flags " in *" $deplib "*) ;; * ) func_append new_inherited_linker_flags " $deplib" ;; esac fi fi continue ;; -L*) case $linkmode in lib) deplibs="$deplib $deplibs" test conv = "$pass" && continue newdependency_libs="$deplib $newdependency_libs" func_stripname '-L' '' "$deplib" func_resolve_sysroot "$func_stripname_result" func_append newlib_search_path " $func_resolve_sysroot_result" ;; prog) if test conv = "$pass"; then deplibs="$deplib $deplibs" continue fi if test scan = "$pass"; then deplibs="$deplib $deplibs" else compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" fi func_stripname '-L' '' "$deplib" func_resolve_sysroot "$func_stripname_result" func_append newlib_search_path " $func_resolve_sysroot_result" ;; *) func_warning "'-L' is ignored for archives/objects" ;; esac # linkmode continue ;; # -L -R*) if test link = "$pass"; then func_stripname '-R' '' "$deplib" func_resolve_sysroot "$func_stripname_result" dir=$func_resolve_sysroot_result # Make sure the xrpath contains only unique directories. case "$xrpath " in *" $dir "*) ;; *) func_append xrpath " $dir" ;; esac fi deplibs="$deplib $deplibs" continue ;; *.la) func_resolve_sysroot "$deplib" lib=$func_resolve_sysroot_result ;; *.$libext) if test conv = "$pass"; then deplibs="$deplib $deplibs" continue fi case $linkmode in lib) # Linking convenience modules into shared libraries is allowed, # but linking other static libraries is non-portable. case " $dlpreconveniencelibs " in *" $deplib "*) ;; *) valid_a_lib=false case $deplibs_check_method in match_pattern*) set dummy $deplibs_check_method; shift match_pattern_regex=`expr "$deplibs_check_method" : "$1 \(.*\)"` if eval "\$ECHO \"$deplib\"" 2>/dev/null | $SED 10q \ | $EGREP "$match_pattern_regex" > /dev/null; then valid_a_lib=: fi ;; pass_all) valid_a_lib=: ;; esac if $valid_a_lib; then echo $ECHO "*** Warning: Linking the shared library $output against the" $ECHO "*** static library $deplib is not portable!" deplibs="$deplib $deplibs" else echo $ECHO "*** Warning: Trying to link with static lib archive $deplib." echo "*** I have the capability to make that library automatically link in when" echo "*** you link to this library. But I can only do this if you have a" echo "*** shared version of the library, which you do not appear to have" echo "*** because the file extensions .$libext of this argument makes me believe" echo "*** that it is just a static archive that I should not use here." fi ;; esac continue ;; prog) if test link != "$pass"; then deplibs="$deplib $deplibs" else compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" fi continue ;; esac # linkmode ;; # *.$libext *.lo | *.$objext) if test conv = "$pass"; then deplibs="$deplib $deplibs" elif test prog = "$linkmode"; then if test dlpreopen = "$pass" || test yes != "$dlopen_support" || test no = "$build_libtool_libs"; then # If there is no dlopen support or we're linking statically, # we need to preload. func_append newdlprefiles " $deplib" compile_deplibs="$deplib $compile_deplibs" finalize_deplibs="$deplib $finalize_deplibs" else func_append newdlfiles " $deplib" fi fi continue ;; %DEPLIBS%) alldeplibs=: continue ;; esac # case $deplib $found || test -f "$lib" \ || func_fatal_error "cannot find the library '$lib' or unhandled argument '$deplib'" # Check to see that this really is a libtool archive. func_lalib_unsafe_p "$lib" \ || func_fatal_error "'$lib' is not a valid libtool archive" func_dirname "$lib" "" "." ladir=$func_dirname_result dlname= dlopen= dlpreopen= libdir= library_names= old_library= inherited_linker_flags= # If the library was installed with an old release of libtool, # it will not redefine variables installed, or shouldnotlink installed=yes shouldnotlink=no avoidtemprpath= # Read the .la file func_source "$lib" # Convert "-framework foo" to "foo.ltframework" if test -n "$inherited_linker_flags"; then tmp_inherited_linker_flags=`$ECHO "$inherited_linker_flags" | $SED 's/-framework \([^ $]*\)/\1.ltframework/g'` for tmp_inherited_linker_flag in $tmp_inherited_linker_flags; do case " $new_inherited_linker_flags " in *" $tmp_inherited_linker_flag "*) ;; *) func_append new_inherited_linker_flags " $tmp_inherited_linker_flag";; esac done fi dependency_libs=`$ECHO " $dependency_libs" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` if test lib,link = "$linkmode,$pass" || test prog,scan = "$linkmode,$pass" || { test prog != "$linkmode" && test lib != "$linkmode"; }; then test -n "$dlopen" && func_append dlfiles " $dlopen" test -n "$dlpreopen" && func_append dlprefiles " $dlpreopen" fi if test conv = "$pass"; then # Only check for convenience libraries deplibs="$lib $deplibs" if test -z "$libdir"; then if test -z "$old_library"; then func_fatal_error "cannot find name of link library for '$lib'" fi # It is a libtool convenience library, so add in its objects. func_append convenience " $ladir/$objdir/$old_library" func_append old_convenience " $ladir/$objdir/$old_library" tmp_libs= for deplib in $dependency_libs; do deplibs="$deplib $deplibs" if $opt_preserve_dup_deps; then case "$tmp_libs " in *" $deplib "*) func_append specialdeplibs " $deplib" ;; esac fi func_append tmp_libs " $deplib" done elif test prog != "$linkmode" && test lib != "$linkmode"; then func_fatal_error "'$lib' is not a convenience library" fi continue fi # $pass = conv # Get the name of the library we link against. linklib= if test -n "$old_library" && { test yes = "$prefer_static_libs" || test built,no = "$prefer_static_libs,$installed"; }; then linklib=$old_library else for l in $old_library $library_names; do linklib=$l done fi if test -z "$linklib"; then func_fatal_error "cannot find name of link library for '$lib'" fi # This library was specified with -dlopen. if test dlopen = "$pass"; then test -z "$libdir" \ && func_fatal_error "cannot -dlopen a convenience library: '$lib'" if test -z "$dlname" || test yes != "$dlopen_support" || test no = "$build_libtool_libs" then # If there is no dlname, no dlopen support or we're linking # statically, we need to preload. We also need to preload any # dependent libraries so libltdl's deplib preloader doesn't # bomb out in the load deplibs phase. func_append dlprefiles " $lib $dependency_libs" else func_append newdlfiles " $lib" fi continue fi # $pass = dlopen # We need an absolute path. case $ladir in [\\/]* | [A-Za-z]:[\\/]*) abs_ladir=$ladir ;; *) abs_ladir=`cd "$ladir" && pwd` if test -z "$abs_ladir"; then func_warning "cannot determine absolute directory name of '$ladir'" func_warning "passing it literally to the linker, although it might fail" abs_ladir=$ladir fi ;; esac func_basename "$lib" laname=$func_basename_result # Find the relevant object directory and library name. if test yes = "$installed"; then if test ! -f "$lt_sysroot$libdir/$linklib" && test -f "$abs_ladir/$linklib"; then func_warning "library '$lib' was moved." dir=$ladir absdir=$abs_ladir libdir=$abs_ladir else dir=$lt_sysroot$libdir absdir=$lt_sysroot$libdir fi test yes = "$hardcode_automatic" && avoidtemprpath=yes else if test ! -f "$ladir/$objdir/$linklib" && test -f "$abs_ladir/$linklib"; then dir=$ladir absdir=$abs_ladir # Remove this search path later func_append notinst_path " $abs_ladir" else dir=$ladir/$objdir absdir=$abs_ladir/$objdir # Remove this search path later func_append notinst_path " $abs_ladir" fi fi # $installed = yes func_stripname 'lib' '.la' "$laname" name=$func_stripname_result # This library was specified with -dlpreopen. if test dlpreopen = "$pass"; then if test -z "$libdir" && test prog = "$linkmode"; then func_fatal_error "only libraries may -dlpreopen a convenience library: '$lib'" fi case $host in # special handling for platforms with PE-DLLs. *cygwin* | *mingw* | *cegcc* ) # Linker will automatically link against shared library if both # static and shared are present. Therefore, ensure we extract # symbols from the import library if a shared library is present # (otherwise, the dlopen module name will be incorrect). We do # this by putting the import library name into $newdlprefiles. # We recover the dlopen module name by 'saving' the la file # name in a special purpose variable, and (later) extracting the # dlname from the la file. if test -n "$dlname"; then func_tr_sh "$dir/$linklib" eval "libfile_$func_tr_sh_result=\$abs_ladir/\$laname" func_append newdlprefiles " $dir/$linklib" else func_append newdlprefiles " $dir/$old_library" # Keep a list of preopened convenience libraries to check # that they are being used correctly in the link pass. test -z "$libdir" && \ func_append dlpreconveniencelibs " $dir/$old_library" fi ;; * ) # Prefer using a static library (so that no silly _DYNAMIC symbols # are required to link). if test -n "$old_library"; then func_append newdlprefiles " $dir/$old_library" # Keep a list of preopened convenience libraries to check # that they are being used correctly in the link pass. test -z "$libdir" && \ func_append dlpreconveniencelibs " $dir/$old_library" # Otherwise, use the dlname, so that lt_dlopen finds it. elif test -n "$dlname"; then func_append newdlprefiles " $dir/$dlname" else func_append newdlprefiles " $dir/$linklib" fi ;; esac fi # $pass = dlpreopen if test -z "$libdir"; then # Link the convenience library if test lib = "$linkmode"; then deplibs="$dir/$old_library $deplibs" elif test prog,link = "$linkmode,$pass"; then compile_deplibs="$dir/$old_library $compile_deplibs" finalize_deplibs="$dir/$old_library $finalize_deplibs" else deplibs="$lib $deplibs" # used for prog,scan pass fi continue fi if test prog = "$linkmode" && test link != "$pass"; then func_append newlib_search_path " $ladir" deplibs="$lib $deplibs" linkalldeplibs=false if test no != "$link_all_deplibs" || test -z "$library_names" || test no = "$build_libtool_libs"; then linkalldeplibs=: fi tmp_libs= for deplib in $dependency_libs; do case $deplib in -L*) func_stripname '-L' '' "$deplib" func_resolve_sysroot "$func_stripname_result" func_append newlib_search_path " $func_resolve_sysroot_result" ;; esac # Need to link against all dependency_libs? if $linkalldeplibs; then deplibs="$deplib $deplibs" else # Need to hardcode shared library paths # or/and link against static libraries newdependency_libs="$deplib $newdependency_libs" fi if $opt_preserve_dup_deps; then case "$tmp_libs " in *" $deplib "*) func_append specialdeplibs " $deplib" ;; esac fi func_append tmp_libs " $deplib" done # for deplib continue fi # $linkmode = prog... if test prog,link = "$linkmode,$pass"; then if test -n "$library_names" && { { test no = "$prefer_static_libs" || test built,yes = "$prefer_static_libs,$installed"; } || test -z "$old_library"; }; then # We need to hardcode the library path if test -n "$shlibpath_var" && test -z "$avoidtemprpath"; then # Make sure the rpath contains only unique directories. case $temp_rpath: in *"$absdir:"*) ;; *) func_append temp_rpath "$absdir:" ;; esac fi # Hardcode the library path. # Skip directories that are in the system default run-time # search path. case " $sys_lib_dlsearch_path " in *" $absdir "*) ;; *) case "$compile_rpath " in *" $absdir "*) ;; *) func_append compile_rpath " $absdir" ;; esac ;; esac case " $sys_lib_dlsearch_path " in *" $libdir "*) ;; *) case "$finalize_rpath " in *" $libdir "*) ;; *) func_append finalize_rpath " $libdir" ;; esac ;; esac fi # $linkmode,$pass = prog,link... if $alldeplibs && { test pass_all = "$deplibs_check_method" || { test yes = "$build_libtool_libs" && test -n "$library_names"; }; }; then # We only need to search for static libraries continue fi fi link_static=no # Whether the deplib will be linked statically use_static_libs=$prefer_static_libs if test built = "$use_static_libs" && test yes = "$installed"; then use_static_libs=no fi if test -n "$library_names" && { test no = "$use_static_libs" || test -z "$old_library"; }; then case $host in *cygwin* | *mingw* | *cegcc* | *os2*) # No point in relinking DLLs because paths are not encoded func_append notinst_deplibs " $lib" need_relink=no ;; *) if test no = "$installed"; then func_append notinst_deplibs " $lib" need_relink=yes fi ;; esac # This is a shared library # Warn about portability, can't link against -module's on some # systems (darwin). Don't bleat about dlopened modules though! dlopenmodule= for dlpremoduletest in $dlprefiles; do if test "X$dlpremoduletest" = "X$lib"; then dlopenmodule=$dlpremoduletest break fi done if test -z "$dlopenmodule" && test yes = "$shouldnotlink" && test link = "$pass"; then echo if test prog = "$linkmode"; then $ECHO "*** Warning: Linking the executable $output against the loadable module" else $ECHO "*** Warning: Linking the shared library $output against the loadable module" fi $ECHO "*** $linklib is not portable!" fi if test lib = "$linkmode" && test yes = "$hardcode_into_libs"; then # Hardcode the library path. # Skip directories that are in the system default run-time # search path. case " $sys_lib_dlsearch_path " in *" $absdir "*) ;; *) case "$compile_rpath " in *" $absdir "*) ;; *) func_append compile_rpath " $absdir" ;; esac ;; esac case " $sys_lib_dlsearch_path " in *" $libdir "*) ;; *) case "$finalize_rpath " in *" $libdir "*) ;; *) func_append finalize_rpath " $libdir" ;; esac ;; esac fi if test -n "$old_archive_from_expsyms_cmds"; then # figure out the soname set dummy $library_names shift realname=$1 shift libname=`eval "\\$ECHO \"$libname_spec\""` # use dlname if we got it. it's perfectly good, no? if test -n "$dlname"; then soname=$dlname elif test -n "$soname_spec"; then # bleh windows case $host in *cygwin* | mingw* | *cegcc* | *os2*) func_arith $current - $age major=$func_arith_result versuffix=-$major ;; esac eval soname=\"$soname_spec\" else soname=$realname fi # Make a new name for the extract_expsyms_cmds to use soroot=$soname func_basename "$soroot" soname=$func_basename_result func_stripname 'lib' '.dll' "$soname" newlib=libimp-$func_stripname_result.a # If the library has no export list, then create one now if test -f "$output_objdir/$soname-def"; then : else func_verbose "extracting exported symbol list from '$soname'" func_execute_cmds "$extract_expsyms_cmds" 'exit $?' fi # Create $newlib if test -f "$output_objdir/$newlib"; then :; else func_verbose "generating import library for '$soname'" func_execute_cmds "$old_archive_from_expsyms_cmds" 'exit $?' fi # make sure the library variables are pointing to the new library dir=$output_objdir linklib=$newlib fi # test -n "$old_archive_from_expsyms_cmds" if test prog = "$linkmode" || test relink != "$opt_mode"; then add_shlibpath= add_dir= add= lib_linked=yes case $hardcode_action in immediate | unsupported) if test no = "$hardcode_direct"; then add=$dir/$linklib case $host in *-*-sco3.2v5.0.[024]*) add_dir=-L$dir ;; *-*-sysv4*uw2*) add_dir=-L$dir ;; *-*-sysv5OpenUNIX* | *-*-sysv5UnixWare7.[01].[10]* | \ *-*-unixware7*) add_dir=-L$dir ;; *-*-darwin* ) # if the lib is a (non-dlopened) module then we cannot # link against it, someone is ignoring the earlier warnings if /usr/bin/file -L $add 2> /dev/null | $GREP ": [^:]* bundle" >/dev/null; then if test "X$dlopenmodule" != "X$lib"; then $ECHO "*** Warning: lib $linklib is a module, not a shared library" if test -z "$old_library"; then echo echo "*** And there doesn't seem to be a static archive available" echo "*** The link will probably fail, sorry" else add=$dir/$old_library fi elif test -n "$old_library"; then add=$dir/$old_library fi fi esac elif test no = "$hardcode_minus_L"; then case $host in *-*-sunos*) add_shlibpath=$dir ;; esac add_dir=-L$dir add=-l$name elif test no = "$hardcode_shlibpath_var"; then add_shlibpath=$dir add=-l$name else lib_linked=no fi ;; relink) if test yes = "$hardcode_direct" && test no = "$hardcode_direct_absolute"; then add=$dir/$linklib elif test yes = "$hardcode_minus_L"; then add_dir=-L$absdir # Try looking first in the location we're being installed to. if test -n "$inst_prefix_dir"; then case $libdir in [\\/]*) func_append add_dir " -L$inst_prefix_dir$libdir" ;; esac fi add=-l$name elif test yes = "$hardcode_shlibpath_var"; then add_shlibpath=$dir add=-l$name else lib_linked=no fi ;; *) lib_linked=no ;; esac if test yes != "$lib_linked"; then func_fatal_configuration "unsupported hardcode properties" fi if test -n "$add_shlibpath"; then case :$compile_shlibpath: in *":$add_shlibpath:"*) ;; *) func_append compile_shlibpath "$add_shlibpath:" ;; esac fi if test prog = "$linkmode"; then test -n "$add_dir" && compile_deplibs="$add_dir $compile_deplibs" test -n "$add" && compile_deplibs="$add $compile_deplibs" else test -n "$add_dir" && deplibs="$add_dir $deplibs" test -n "$add" && deplibs="$add $deplibs" if test yes != "$hardcode_direct" && test yes != "$hardcode_minus_L" && test yes = "$hardcode_shlibpath_var"; then case :$finalize_shlibpath: in *":$libdir:"*) ;; *) func_append finalize_shlibpath "$libdir:" ;; esac fi fi fi if test prog = "$linkmode" || test relink = "$opt_mode"; then add_shlibpath= add_dir= add= # Finalize command for both is simple: just hardcode it. if test yes = "$hardcode_direct" && test no = "$hardcode_direct_absolute"; then add=$libdir/$linklib elif test yes = "$hardcode_minus_L"; then add_dir=-L$libdir add=-l$name elif test yes = "$hardcode_shlibpath_var"; then case :$finalize_shlibpath: in *":$libdir:"*) ;; *) func_append finalize_shlibpath "$libdir:" ;; esac add=-l$name elif test yes = "$hardcode_automatic"; then if test -n "$inst_prefix_dir" && test -f "$inst_prefix_dir$libdir/$linklib"; then add=$inst_prefix_dir$libdir/$linklib else add=$libdir/$linklib fi else # We cannot seem to hardcode it, guess we'll fake it. add_dir=-L$libdir # Try looking first in the location we're being installed to. if test -n "$inst_prefix_dir"; then case $libdir in [\\/]*) func_append add_dir " -L$inst_prefix_dir$libdir" ;; esac fi add=-l$name fi if test prog = "$linkmode"; then test -n "$add_dir" && finalize_deplibs="$add_dir $finalize_deplibs" test -n "$add" && finalize_deplibs="$add $finalize_deplibs" else test -n "$add_dir" && deplibs="$add_dir $deplibs" test -n "$add" && deplibs="$add $deplibs" fi fi elif test prog = "$linkmode"; then # Here we assume that one of hardcode_direct or hardcode_minus_L # is not unsupported. This is valid on all known static and # shared platforms. if test unsupported != "$hardcode_direct"; then test -n "$old_library" && linklib=$old_library compile_deplibs="$dir/$linklib $compile_deplibs" finalize_deplibs="$dir/$linklib $finalize_deplibs" else compile_deplibs="-l$name -L$dir $compile_deplibs" finalize_deplibs="-l$name -L$dir $finalize_deplibs" fi elif test yes = "$build_libtool_libs"; then # Not a shared library if test pass_all != "$deplibs_check_method"; then # We're trying link a shared library against a static one # but the system doesn't support it. # Just print a warning and add the library to dependency_libs so # that the program can be linked against the static library. echo $ECHO "*** Warning: This system cannot link to static lib archive $lib." echo "*** I have the capability to make that library automatically link in when" echo "*** you link to this library. But I can only do this if you have a" echo "*** shared version of the library, which you do not appear to have." if test yes = "$module"; then echo "*** But as you try to build a module library, libtool will still create " echo "*** a static module, that should work as long as the dlopening application" echo "*** is linked with the -dlopen flag to resolve symbols at runtime." if test -z "$global_symbol_pipe"; then echo echo "*** However, this would only work if libtool was able to extract symbol" echo "*** lists from a program, using 'nm' or equivalent, but libtool could" echo "*** not find such a program. So, this module is probably useless." echo "*** 'nm' from GNU binutils and a full rebuild may help." fi if test no = "$build_old_libs"; then build_libtool_libs=module build_old_libs=yes else build_libtool_libs=no fi fi else deplibs="$dir/$old_library $deplibs" link_static=yes fi fi # link shared/static library? if test lib = "$linkmode"; then if test -n "$dependency_libs" && { test yes != "$hardcode_into_libs" || test yes = "$build_old_libs" || test yes = "$link_static"; }; then # Extract -R from dependency_libs temp_deplibs= for libdir in $dependency_libs; do case $libdir in -R*) func_stripname '-R' '' "$libdir" temp_xrpath=$func_stripname_result case " $xrpath " in *" $temp_xrpath "*) ;; *) func_append xrpath " $temp_xrpath";; esac;; *) func_append temp_deplibs " $libdir";; esac done dependency_libs=$temp_deplibs fi func_append newlib_search_path " $absdir" # Link against this library test no = "$link_static" && newdependency_libs="$abs_ladir/$laname $newdependency_libs" # ... and its dependency_libs tmp_libs= for deplib in $dependency_libs; do newdependency_libs="$deplib $newdependency_libs" case $deplib in -L*) func_stripname '-L' '' "$deplib" func_resolve_sysroot "$func_stripname_result";; *) func_resolve_sysroot "$deplib" ;; esac if $opt_preserve_dup_deps; then case "$tmp_libs " in *" $func_resolve_sysroot_result "*) func_append specialdeplibs " $func_resolve_sysroot_result" ;; esac fi func_append tmp_libs " $func_resolve_sysroot_result" done if test no != "$link_all_deplibs"; then # Add the search paths of all dependency libraries for deplib in $dependency_libs; do path= case $deplib in -L*) path=$deplib ;; *.la) func_resolve_sysroot "$deplib" deplib=$func_resolve_sysroot_result func_dirname "$deplib" "" "." dir=$func_dirname_result # We need an absolute path. case $dir in [\\/]* | [A-Za-z]:[\\/]*) absdir=$dir ;; *) absdir=`cd "$dir" && pwd` if test -z "$absdir"; then func_warning "cannot determine absolute directory name of '$dir'" absdir=$dir fi ;; esac if $GREP "^installed=no" $deplib > /dev/null; then case $host in *-*-darwin*) depdepl= eval deplibrary_names=`$SED -n -e 's/^library_names=\(.*\)$/\1/p' $deplib` if test -n "$deplibrary_names"; then for tmp in $deplibrary_names; do depdepl=$tmp done if test -f "$absdir/$objdir/$depdepl"; then depdepl=$absdir/$objdir/$depdepl darwin_install_name=`$OTOOL -L $depdepl | awk '{if (NR == 2) {print $1;exit}}'` if test -z "$darwin_install_name"; then darwin_install_name=`$OTOOL64 -L $depdepl | awk '{if (NR == 2) {print $1;exit}}'` fi func_append compiler_flags " $wl-dylib_file $wl$darwin_install_name:$depdepl" func_append linker_flags " -dylib_file $darwin_install_name:$depdepl" path= fi fi ;; *) path=-L$absdir/$objdir ;; esac else eval libdir=`$SED -n -e 's/^libdir=\(.*\)$/\1/p' $deplib` test -z "$libdir" && \ func_fatal_error "'$deplib' is not a valid libtool archive" test "$absdir" != "$libdir" && \ func_warning "'$deplib' seems to be moved" path=-L$absdir fi ;; esac case " $deplibs " in *" $path "*) ;; *) deplibs="$path $deplibs" ;; esac done fi # link_all_deplibs != no fi # linkmode = lib done # for deplib in $libs if test link = "$pass"; then if test prog = "$linkmode"; then compile_deplibs="$new_inherited_linker_flags $compile_deplibs" finalize_deplibs="$new_inherited_linker_flags $finalize_deplibs" else compiler_flags="$compiler_flags "`$ECHO " $new_inherited_linker_flags" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` fi fi dependency_libs=$newdependency_libs if test dlpreopen = "$pass"; then # Link the dlpreopened libraries before other libraries for deplib in $save_deplibs; do deplibs="$deplib $deplibs" done fi if test dlopen != "$pass"; then test conv = "$pass" || { # Make sure lib_search_path contains only unique directories. lib_search_path= for dir in $newlib_search_path; do case "$lib_search_path " in *" $dir "*) ;; *) func_append lib_search_path " $dir" ;; esac done newlib_search_path= } if test prog,link = "$linkmode,$pass"; then vars="compile_deplibs finalize_deplibs" else vars=deplibs fi for var in $vars dependency_libs; do # Add libraries to $var in reverse order eval tmp_libs=\"\$$var\" new_libs= for deplib in $tmp_libs; do # FIXME: Pedantically, this is the right thing to do, so # that some nasty dependency loop isn't accidentally # broken: #new_libs="$deplib $new_libs" # Pragmatically, this seems to cause very few problems in # practice: case $deplib in -L*) new_libs="$deplib $new_libs" ;; -R*) ;; *) # And here is the reason: when a library appears more # than once as an explicit dependence of a library, or # is implicitly linked in more than once by the # compiler, it is considered special, and multiple # occurrences thereof are not removed. Compare this # with having the same library being listed as a # dependency of multiple other libraries: in this case, # we know (pedantically, we assume) the library does not # need to be listed more than once, so we keep only the # last copy. This is not always right, but it is rare # enough that we require users that really mean to play # such unportable linking tricks to link the library # using -Wl,-lname, so that libtool does not consider it # for duplicate removal. case " $specialdeplibs " in *" $deplib "*) new_libs="$deplib $new_libs" ;; *) case " $new_libs " in *" $deplib "*) ;; *) new_libs="$deplib $new_libs" ;; esac ;; esac ;; esac done tmp_libs= for deplib in $new_libs; do case $deplib in -L*) case " $tmp_libs " in *" $deplib "*) ;; *) func_append tmp_libs " $deplib" ;; esac ;; *) func_append tmp_libs " $deplib" ;; esac done eval $var=\"$tmp_libs\" done # for var fi # Add Sun CC postdeps if required: test CXX = "$tagname" && { case $host_os in linux*) case `$CC -V 2>&1 | sed 5q` in *Sun\ C*) # Sun C++ 5.9 func_suncc_cstd_abi if test no != "$suncc_use_cstd_abi"; then func_append postdeps ' -library=Cstd -library=Crun' fi ;; esac ;; solaris*) func_cc_basename "$CC" case $func_cc_basename_result in CC* | sunCC*) func_suncc_cstd_abi if test no != "$suncc_use_cstd_abi"; then func_append postdeps ' -library=Cstd -library=Crun' fi ;; esac ;; esac } # Last step: remove runtime libs from dependency_libs # (they stay in deplibs) tmp_libs= for i in $dependency_libs; do case " $predeps $postdeps $compiler_lib_search_path " in *" $i "*) i= ;; esac if test -n "$i"; then func_append tmp_libs " $i" fi done dependency_libs=$tmp_libs done # for pass if test prog = "$linkmode"; then dlfiles=$newdlfiles fi if test prog = "$linkmode" || test lib = "$linkmode"; then dlprefiles=$newdlprefiles fi case $linkmode in oldlib) if test -n "$dlfiles$dlprefiles" || test no != "$dlself"; then func_warning "'-dlopen' is ignored for archives" fi case " $deplibs" in *\ -l* | *\ -L*) func_warning "'-l' and '-L' are ignored for archives" ;; esac test -n "$rpath" && \ func_warning "'-rpath' is ignored for archives" test -n "$xrpath" && \ func_warning "'-R' is ignored for archives" test -n "$vinfo" && \ func_warning "'-version-info/-version-number' is ignored for archives" test -n "$release" && \ func_warning "'-release' is ignored for archives" test -n "$export_symbols$export_symbols_regex" && \ func_warning "'-export-symbols' is ignored for archives" # Now set the variables for building old libraries. build_libtool_libs=no oldlibs=$output func_append objs "$old_deplibs" ;; lib) # Make sure we only generate libraries of the form 'libNAME.la'. case $outputname in lib*) func_stripname 'lib' '.la' "$outputname" name=$func_stripname_result eval shared_ext=\"$shrext_cmds\" eval libname=\"$libname_spec\" ;; *) test no = "$module" \ && func_fatal_help "libtool library '$output' must begin with 'lib'" if test no != "$need_lib_prefix"; then # Add the "lib" prefix for modules if required func_stripname '' '.la' "$outputname" name=$func_stripname_result eval shared_ext=\"$shrext_cmds\" eval libname=\"$libname_spec\" else func_stripname '' '.la' "$outputname" libname=$func_stripname_result fi ;; esac if test -n "$objs"; then if test pass_all != "$deplibs_check_method"; then func_fatal_error "cannot build libtool library '$output' from non-libtool objects on this host:$objs" else echo $ECHO "*** Warning: Linking the shared library $output against the non-libtool" $ECHO "*** objects $objs is not portable!" func_append libobjs " $objs" fi fi test no = "$dlself" \ || func_warning "'-dlopen self' is ignored for libtool libraries" set dummy $rpath shift test 1 -lt "$#" \ && func_warning "ignoring multiple '-rpath's for a libtool library" install_libdir=$1 oldlibs= if test -z "$rpath"; then if test yes = "$build_libtool_libs"; then # Building a libtool convenience library. # Some compilers have problems with a '.al' extension so # convenience libraries should have the same extension an # archive normally would. oldlibs="$output_objdir/$libname.$libext $oldlibs" build_libtool_libs=convenience build_old_libs=yes fi test -n "$vinfo" && \ func_warning "'-version-info/-version-number' is ignored for convenience libraries" test -n "$release" && \ func_warning "'-release' is ignored for convenience libraries" else # Parse the version information argument. save_ifs=$IFS; IFS=: set dummy $vinfo 0 0 0 shift IFS=$save_ifs test -n "$7" && \ func_fatal_help "too many parameters to '-version-info'" # convert absolute version numbers to libtool ages # this retains compatibility with .la files and attempts # to make the code below a bit more comprehensible case $vinfo_number in yes) number_major=$1 number_minor=$2 number_revision=$3 # # There are really only two kinds -- those that # use the current revision as the major version # and those that subtract age and use age as # a minor version. But, then there is irix # that has an extra 1 added just for fun # case $version_type in # correct linux to gnu/linux during the next big refactor darwin|freebsd-elf|linux|osf|windows|none) func_arith $number_major + $number_minor current=$func_arith_result age=$number_minor revision=$number_revision ;; freebsd-aout|qnx|sunos) current=$number_major revision=$number_minor age=0 ;; irix|nonstopux) func_arith $number_major + $number_minor current=$func_arith_result age=$number_minor revision=$number_minor lt_irix_increment=no ;; *) func_fatal_configuration "$modename: unknown library version type '$version_type'" ;; esac ;; no) current=$1 revision=$2 age=$3 ;; esac # Check that each of the things are valid numbers. case $current in 0|[1-9]|[1-9][0-9]|[1-9][0-9][0-9]|[1-9][0-9][0-9][0-9]|[1-9][0-9][0-9][0-9][0-9]) ;; *) func_error "CURRENT '$current' must be a nonnegative integer" func_fatal_error "'$vinfo' is not valid version information" ;; esac case $revision in 0|[1-9]|[1-9][0-9]|[1-9][0-9][0-9]|[1-9][0-9][0-9][0-9]|[1-9][0-9][0-9][0-9][0-9]) ;; *) func_error "REVISION '$revision' must be a nonnegative integer" func_fatal_error "'$vinfo' is not valid version information" ;; esac case $age in 0|[1-9]|[1-9][0-9]|[1-9][0-9][0-9]|[1-9][0-9][0-9][0-9]|[1-9][0-9][0-9][0-9][0-9]) ;; *) func_error "AGE '$age' must be a nonnegative integer" func_fatal_error "'$vinfo' is not valid version information" ;; esac if test "$age" -gt "$current"; then func_error "AGE '$age' is greater than the current interface number '$current'" func_fatal_error "'$vinfo' is not valid version information" fi # Calculate the version variables. major= versuffix= verstring= case $version_type in none) ;; darwin) # Like Linux, but with the current version available in # verstring for coding it into the library header func_arith $current - $age major=.$func_arith_result versuffix=$major.$age.$revision # Darwin ld doesn't like 0 for these options... func_arith $current + 1 minor_current=$func_arith_result xlcverstring="$wl-compatibility_version $wl$minor_current $wl-current_version $wl$minor_current.$revision" verstring="-compatibility_version $minor_current -current_version $minor_current.$revision" # On Darwin other compilers case $CC in nagfor*) verstring="$wl-compatibility_version $wl$minor_current $wl-current_version $wl$minor_current.$revision" ;; *) verstring="-compatibility_version $minor_current -current_version $minor_current.$revision" ;; esac ;; freebsd-aout) major=.$current versuffix=.$current.$revision ;; freebsd-elf) func_arith $current - $age major=.$func_arith_result versuffix=$major.$age.$revision ;; irix | nonstopux) if test no = "$lt_irix_increment"; then func_arith $current - $age else func_arith $current - $age + 1 fi major=$func_arith_result case $version_type in nonstopux) verstring_prefix=nonstopux ;; *) verstring_prefix=sgi ;; esac verstring=$verstring_prefix$major.$revision # Add in all the interfaces that we are compatible with. loop=$revision while test 0 -ne "$loop"; do func_arith $revision - $loop iface=$func_arith_result func_arith $loop - 1 loop=$func_arith_result verstring=$verstring_prefix$major.$iface:$verstring done # Before this point, $major must not contain '.'. major=.$major versuffix=$major.$revision ;; linux) # correct to gnu/linux during the next big refactor func_arith $current - $age major=.$func_arith_result versuffix=$major.$age.$revision ;; osf) func_arith $current - $age major=.$func_arith_result versuffix=.$current.$age.$revision verstring=$current.$age.$revision # Add in all the interfaces that we are compatible with. loop=$age while test 0 -ne "$loop"; do func_arith $current - $loop iface=$func_arith_result func_arith $loop - 1 loop=$func_arith_result verstring=$verstring:$iface.0 done # Make executables depend on our current version. func_append verstring ":$current.0" ;; qnx) major=.$current versuffix=.$current ;; sco) major=.$current versuffix=.$current ;; sunos) major=.$current versuffix=.$current.$revision ;; windows) # Use '-' rather than '.', since we only want one # extension on DOS 8.3 file systems. func_arith $current - $age major=$func_arith_result versuffix=-$major ;; *) func_fatal_configuration "unknown library version type '$version_type'" ;; esac # Clear the version info if we defaulted, and they specified a release. if test -z "$vinfo" && test -n "$release"; then major= case $version_type in darwin) # we can't check for "0.0" in archive_cmds due to quoting # problems, so we reset it completely verstring= ;; *) verstring=0.0 ;; esac if test no = "$need_version"; then versuffix= else versuffix=.0.0 fi fi # Remove version info from name if versioning should be avoided if test yes,no = "$avoid_version,$need_version"; then major= versuffix= verstring= fi # Check to see if the archive will have undefined symbols. if test yes = "$allow_undefined"; then if test unsupported = "$allow_undefined_flag"; then if test yes = "$build_old_libs"; then func_warning "undefined symbols not allowed in $host shared libraries; building static only" build_libtool_libs=no else func_fatal_error "can't build $host shared library unless -no-undefined is specified" fi fi else # Don't allow undefined symbols. allow_undefined_flag=$no_undefined_flag fi fi func_generate_dlsyms "$libname" "$libname" : func_append libobjs " $symfileobj" test " " = "$libobjs" && libobjs= if test relink != "$opt_mode"; then # Remove our outputs, but don't remove object files since they # may have been created when compiling PIC objects. removelist= tempremovelist=`$ECHO "$output_objdir/*"` for p in $tempremovelist; do case $p in *.$objext | *.gcno) ;; $output_objdir/$outputname | $output_objdir/$libname.* | $output_objdir/$libname$release.*) if test -n "$precious_files_regex"; then if $ECHO "$p" | $EGREP -e "$precious_files_regex" >/dev/null 2>&1 then continue fi fi func_append removelist " $p" ;; *) ;; esac done test -n "$removelist" && \ func_show_eval "${RM}r \$removelist" fi # Now set the variables for building old libraries. if test yes = "$build_old_libs" && test convenience != "$build_libtool_libs"; then func_append oldlibs " $output_objdir/$libname.$libext" # Transform .lo files to .o files. oldobjs="$objs "`$ECHO "$libobjs" | $SP2NL | $SED "/\.$libext$/d; $lo2o" | $NL2SP` fi # Eliminate all temporary directories. #for path in $notinst_path; do # lib_search_path=`$ECHO "$lib_search_path " | $SED "s% $path % %g"` # deplibs=`$ECHO "$deplibs " | $SED "s% -L$path % %g"` # dependency_libs=`$ECHO "$dependency_libs " | $SED "s% -L$path % %g"` #done if test -n "$xrpath"; then # If the user specified any rpath flags, then add them. temp_xrpath= for libdir in $xrpath; do func_replace_sysroot "$libdir" func_append temp_xrpath " -R$func_replace_sysroot_result" case "$finalize_rpath " in *" $libdir "*) ;; *) func_append finalize_rpath " $libdir" ;; esac done if test yes != "$hardcode_into_libs" || test yes = "$build_old_libs"; then dependency_libs="$temp_xrpath $dependency_libs" fi fi # Make sure dlfiles contains only unique files that won't be dlpreopened old_dlfiles=$dlfiles dlfiles= for lib in $old_dlfiles; do case " $dlprefiles $dlfiles " in *" $lib "*) ;; *) func_append dlfiles " $lib" ;; esac done # Make sure dlprefiles contains only unique files old_dlprefiles=$dlprefiles dlprefiles= for lib in $old_dlprefiles; do case "$dlprefiles " in *" $lib "*) ;; *) func_append dlprefiles " $lib" ;; esac done if test yes = "$build_libtool_libs"; then if test -n "$rpath"; then case $host in *-*-cygwin* | *-*-mingw* | *-*-pw32* | *-*-os2* | *-*-beos* | *-cegcc* | *-*-haiku*) # these systems don't actually have a c library (as such)! ;; *-*-rhapsody* | *-*-darwin1.[012]) # Rhapsody C library is in the System framework func_append deplibs " System.ltframework" ;; *-*-netbsd*) # Don't link with libc until the a.out ld.so is fixed. ;; *-*-openbsd* | *-*-freebsd* | *-*-dragonfly*) # Do not include libc due to us having libc/libc_r. ;; *-*-sco3.2v5* | *-*-sco5v6*) # Causes problems with __ctype ;; *-*-sysv4.2uw2* | *-*-sysv5* | *-*-unixware* | *-*-OpenUNIX*) # Compiler inserts libc in the correct place for threads to work ;; *) # Add libc to deplibs on all other systems if necessary. if test yes = "$build_libtool_need_lc"; then func_append deplibs " -lc" fi ;; esac fi # Transform deplibs into only deplibs that can be linked in shared. name_save=$name libname_save=$libname release_save=$release versuffix_save=$versuffix major_save=$major # I'm not sure if I'm treating the release correctly. I think # release should show up in the -l (ie -lgmp5) so we don't want to # add it in twice. Is that correct? release= versuffix= major= newdeplibs= droppeddeps=no case $deplibs_check_method in pass_all) # Don't check for shared/static. Everything works. # This might be a little naive. We might want to check # whether the library exists or not. But this is on # osf3 & osf4 and I'm not really sure... Just # implementing what was already the behavior. newdeplibs=$deplibs ;; test_compile) # This code stresses the "libraries are programs" paradigm to its # limits. Maybe even breaks it. We compile a program, linking it # against the deplibs as a proxy for the library. Then we can check # whether they linked in statically or dynamically with ldd. $opt_dry_run || $RM conftest.c cat > conftest.c <<EOF int main() { return 0; } EOF $opt_dry_run || $RM conftest if $LTCC $LTCFLAGS -o conftest conftest.c $deplibs; then ldd_output=`ldd conftest` for i in $deplibs; do case $i in -l*) func_stripname -l '' "$i" name=$func_stripname_result if test yes = "$allow_libtool_libs_with_static_runtimes"; then case " $predeps $postdeps " in *" $i "*) func_append newdeplibs " $i" i= ;; esac fi if test -n "$i"; then libname=`eval "\\$ECHO \"$libname_spec\""` deplib_matches=`eval "\\$ECHO \"$library_names_spec\""` set dummy $deplib_matches; shift deplib_match=$1 if test `expr "$ldd_output" : ".*$deplib_match"` -ne 0; then func_append newdeplibs " $i" else droppeddeps=yes echo $ECHO "*** Warning: dynamic linker does not accept needed library $i." echo "*** I have the capability to make that library automatically link in when" echo "*** you link to this library. But I can only do this if you have a" echo "*** shared version of the library, which I believe you do not have" echo "*** because a test_compile did reveal that the linker did not use it for" echo "*** its dynamic dependency list that programs get resolved with at runtime." fi fi ;; *) func_append newdeplibs " $i" ;; esac done else # Error occurred in the first compile. Let's try to salvage # the situation: Compile a separate program for each library. for i in $deplibs; do case $i in -l*) func_stripname -l '' "$i" name=$func_stripname_result $opt_dry_run || $RM conftest if $LTCC $LTCFLAGS -o conftest conftest.c $i; then ldd_output=`ldd conftest` if test yes = "$allow_libtool_libs_with_static_runtimes"; then case " $predeps $postdeps " in *" $i "*) func_append newdeplibs " $i" i= ;; esac fi if test -n "$i"; then libname=`eval "\\$ECHO \"$libname_spec\""` deplib_matches=`eval "\\$ECHO \"$library_names_spec\""` set dummy $deplib_matches; shift deplib_match=$1 if test `expr "$ldd_output" : ".*$deplib_match"` -ne 0; then func_append newdeplibs " $i" else droppeddeps=yes echo $ECHO "*** Warning: dynamic linker does not accept needed library $i." echo "*** I have the capability to make that library automatically link in when" echo "*** you link to this library. But I can only do this if you have a" echo "*** shared version of the library, which you do not appear to have" echo "*** because a test_compile did reveal that the linker did not use this one" echo "*** as a dynamic dependency that programs can get resolved with at runtime." fi fi else droppeddeps=yes echo $ECHO "*** Warning! Library $i is needed by this library but I was not able to" echo "*** make it link in! You will probably need to install it or some" echo "*** library that it depends on before this library will be fully" echo "*** functional. Installing it before continuing would be even better." fi ;; *) func_append newdeplibs " $i" ;; esac done fi ;; file_magic*) set dummy $deplibs_check_method; shift file_magic_regex=`expr "$deplibs_check_method" : "$1 \(.*\)"` for a_deplib in $deplibs; do case $a_deplib in -l*) func_stripname -l '' "$a_deplib" name=$func_stripname_result if test yes = "$allow_libtool_libs_with_static_runtimes"; then case " $predeps $postdeps " in *" $a_deplib "*) func_append newdeplibs " $a_deplib" a_deplib= ;; esac fi if test -n "$a_deplib"; then libname=`eval "\\$ECHO \"$libname_spec\""` if test -n "$file_magic_glob"; then libnameglob=`func_echo_all "$libname" | $SED -e $file_magic_glob` else libnameglob=$libname fi test yes = "$want_nocaseglob" && nocaseglob=`shopt -p nocaseglob` for i in $lib_search_path $sys_lib_search_path $shlib_search_path; do if test yes = "$want_nocaseglob"; then shopt -s nocaseglob potential_libs=`ls $i/$libnameglob[.-]* 2>/dev/null` $nocaseglob else potential_libs=`ls $i/$libnameglob[.-]* 2>/dev/null` fi for potent_lib in $potential_libs; do # Follow soft links. if ls -lLd "$potent_lib" 2>/dev/null | $GREP " -> " >/dev/null; then continue fi # The statement above tries to avoid entering an # endless loop below, in case of cyclic links. # We might still enter an endless loop, since a link # loop can be closed while we follow links, # but so what? potlib=$potent_lib while test -h "$potlib" 2>/dev/null; do potliblink=`ls -ld $potlib | $SED 's/.* -> //'` case $potliblink in [\\/]* | [A-Za-z]:[\\/]*) potlib=$potliblink;; *) potlib=`$ECHO "$potlib" | $SED 's|[^/]*$||'`"$potliblink";; esac done if eval $file_magic_cmd \"\$potlib\" 2>/dev/null | $SED -e 10q | $EGREP "$file_magic_regex" > /dev/null; then func_append newdeplibs " $a_deplib" a_deplib= break 2 fi done done fi if test -n "$a_deplib"; then droppeddeps=yes echo $ECHO "*** Warning: linker path does not have real file for library $a_deplib." echo "*** I have the capability to make that library automatically link in when" echo "*** you link to this library. But I can only do this if you have a" echo "*** shared version of the library, which you do not appear to have" echo "*** because I did check the linker path looking for a file starting" if test -z "$potlib"; then $ECHO "*** with $libname but no candidates were found. (...for file magic test)" else $ECHO "*** with $libname and none of the candidates passed a file format test" $ECHO "*** using a file magic. Last file checked: $potlib" fi fi ;; *) # Add a -L argument. func_append newdeplibs " $a_deplib" ;; esac done # Gone through all deplibs. ;; match_pattern*) set dummy $deplibs_check_method; shift match_pattern_regex=`expr "$deplibs_check_method" : "$1 \(.*\)"` for a_deplib in $deplibs; do case $a_deplib in -l*) func_stripname -l '' "$a_deplib" name=$func_stripname_result if test yes = "$allow_libtool_libs_with_static_runtimes"; then case " $predeps $postdeps " in *" $a_deplib "*) func_append newdeplibs " $a_deplib" a_deplib= ;; esac fi if test -n "$a_deplib"; then libname=`eval "\\$ECHO \"$libname_spec\""` for i in $lib_search_path $sys_lib_search_path $shlib_search_path; do potential_libs=`ls $i/$libname[.-]* 2>/dev/null` for potent_lib in $potential_libs; do potlib=$potent_lib # see symlink-check above in file_magic test if eval "\$ECHO \"$potent_lib\"" 2>/dev/null | $SED 10q | \ $EGREP "$match_pattern_regex" > /dev/null; then func_append newdeplibs " $a_deplib" a_deplib= break 2 fi done done fi if test -n "$a_deplib"; then droppeddeps=yes echo $ECHO "*** Warning: linker path does not have real file for library $a_deplib." echo "*** I have the capability to make that library automatically link in when" echo "*** you link to this library. But I can only do this if you have a" echo "*** shared version of the library, which you do not appear to have" echo "*** because I did check the linker path looking for a file starting" if test -z "$potlib"; then $ECHO "*** with $libname but no candidates were found. (...for regex pattern test)" else $ECHO "*** with $libname and none of the candidates passed a file format test" $ECHO "*** using a regex pattern. Last file checked: $potlib" fi fi ;; *) # Add a -L argument. func_append newdeplibs " $a_deplib" ;; esac done # Gone through all deplibs. ;; none | unknown | *) newdeplibs= tmp_deplibs=`$ECHO " $deplibs" | $SED 's/ -lc$//; s/ -[LR][^ ]*//g'` if test yes = "$allow_libtool_libs_with_static_runtimes"; then for i in $predeps $postdeps; do # can't use Xsed below, because $i might contain '/' tmp_deplibs=`$ECHO " $tmp_deplibs" | $SED "s|$i||"` done fi case $tmp_deplibs in *[!\ \ ]*) echo if test none = "$deplibs_check_method"; then echo "*** Warning: inter-library dependencies are not supported in this platform." else echo "*** Warning: inter-library dependencies are not known to be supported." fi echo "*** All declared inter-library dependencies are being dropped." droppeddeps=yes ;; esac ;; esac versuffix=$versuffix_save major=$major_save release=$release_save libname=$libname_save name=$name_save case $host in *-*-rhapsody* | *-*-darwin1.[012]) # On Rhapsody replace the C library with the System framework newdeplibs=`$ECHO " $newdeplibs" | $SED 's/ -lc / System.ltframework /'` ;; esac if test yes = "$droppeddeps"; then if test yes = "$module"; then echo echo "*** Warning: libtool could not satisfy all declared inter-library" $ECHO "*** dependencies of module $libname. Therefore, libtool will create" echo "*** a static module, that should work as long as the dlopening" echo "*** application is linked with the -dlopen flag." if test -z "$global_symbol_pipe"; then echo echo "*** However, this would only work if libtool was able to extract symbol" echo "*** lists from a program, using 'nm' or equivalent, but libtool could" echo "*** not find such a program. So, this module is probably useless." echo "*** 'nm' from GNU binutils and a full rebuild may help." fi if test no = "$build_old_libs"; then oldlibs=$output_objdir/$libname.$libext build_libtool_libs=module build_old_libs=yes else build_libtool_libs=no fi else echo "*** The inter-library dependencies that have been dropped here will be" echo "*** automatically added whenever a program is linked with this library" echo "*** or is declared to -dlopen it." if test no = "$allow_undefined"; then echo echo "*** Since this library must not contain undefined symbols," echo "*** because either the platform does not support them or" echo "*** it was explicitly requested with -no-undefined," echo "*** libtool will only create a static version of it." if test no = "$build_old_libs"; then oldlibs=$output_objdir/$libname.$libext build_libtool_libs=module build_old_libs=yes else build_libtool_libs=no fi fi fi fi # Done checking deplibs! deplibs=$newdeplibs fi # Time to change all our "foo.ltframework" stuff back to "-framework foo" case $host in *-*-darwin*) newdeplibs=`$ECHO " $newdeplibs" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` new_inherited_linker_flags=`$ECHO " $new_inherited_linker_flags" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` deplibs=`$ECHO " $deplibs" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` ;; esac # move library search paths that coincide with paths to not yet # installed libraries to the beginning of the library search list new_libs= for path in $notinst_path; do case " $new_libs " in *" -L$path/$objdir "*) ;; *) case " $deplibs " in *" -L$path/$objdir "*) func_append new_libs " -L$path/$objdir" ;; esac ;; esac done for deplib in $deplibs; do case $deplib in -L*) case " $new_libs " in *" $deplib "*) ;; *) func_append new_libs " $deplib" ;; esac ;; *) func_append new_libs " $deplib" ;; esac done deplibs=$new_libs # All the library-specific variables (install_libdir is set above). library_names= old_library= dlname= # Test again, we may have decided not to build it any more if test yes = "$build_libtool_libs"; then # Remove $wl instances when linking with ld. # FIXME: should test the right _cmds variable. case $archive_cmds in *\$LD\ *) wl= ;; esac if test yes = "$hardcode_into_libs"; then # Hardcode the library paths hardcode_libdirs= dep_rpath= rpath=$finalize_rpath test relink = "$opt_mode" || rpath=$compile_rpath$rpath for libdir in $rpath; do if test -n "$hardcode_libdir_flag_spec"; then if test -n "$hardcode_libdir_separator"; then func_replace_sysroot "$libdir" libdir=$func_replace_sysroot_result if test -z "$hardcode_libdirs"; then hardcode_libdirs=$libdir else # Just accumulate the unique libdirs. case $hardcode_libdir_separator$hardcode_libdirs$hardcode_libdir_separator in *"$hardcode_libdir_separator$libdir$hardcode_libdir_separator"*) ;; *) func_append hardcode_libdirs "$hardcode_libdir_separator$libdir" ;; esac fi else eval flag=\"$hardcode_libdir_flag_spec\" func_append dep_rpath " $flag" fi elif test -n "$runpath_var"; then case "$perm_rpath " in *" $libdir "*) ;; *) func_append perm_rpath " $libdir" ;; esac fi done # Substitute the hardcoded libdirs into the rpath. if test -n "$hardcode_libdir_separator" && test -n "$hardcode_libdirs"; then libdir=$hardcode_libdirs eval "dep_rpath=\"$hardcode_libdir_flag_spec\"" fi if test -n "$runpath_var" && test -n "$perm_rpath"; then # We should set the runpath_var. rpath= for dir in $perm_rpath; do func_append rpath "$dir:" done eval "$runpath_var='$rpath\$$runpath_var'; export $runpath_var" fi test -n "$dep_rpath" && deplibs="$dep_rpath $deplibs" fi shlibpath=$finalize_shlibpath test relink = "$opt_mode" || shlibpath=$compile_shlibpath$shlibpath if test -n "$shlibpath"; then eval "$shlibpath_var='$shlibpath\$$shlibpath_var'; export $shlibpath_var" fi # Get the real and link names of the library. eval shared_ext=\"$shrext_cmds\" eval library_names=\"$library_names_spec\" set dummy $library_names shift realname=$1 shift if test -n "$soname_spec"; then eval soname=\"$soname_spec\" else soname=$realname fi if test -z "$dlname"; then dlname=$soname fi lib=$output_objdir/$realname linknames= for link do func_append linknames " $link" done # Use standard objects if they are pic test -z "$pic_flag" && libobjs=`$ECHO "$libobjs" | $SP2NL | $SED "$lo2o" | $NL2SP` test "X$libobjs" = "X " && libobjs= delfiles= if test -n "$export_symbols" && test -n "$include_expsyms"; then $opt_dry_run || cp "$export_symbols" "$output_objdir/$libname.uexp" export_symbols=$output_objdir/$libname.uexp func_append delfiles " $export_symbols" fi orig_export_symbols= case $host_os in cygwin* | mingw* | cegcc*) if test -n "$export_symbols" && test -z "$export_symbols_regex"; then # exporting using user supplied symfile func_dll_def_p "$export_symbols" || { # and it's NOT already a .def file. Must figure out # which of the given symbols are data symbols and tag # them as such. So, trigger use of export_symbols_cmds. # export_symbols gets reassigned inside the "prepare # the list of exported symbols" if statement, so the # include_expsyms logic still works. orig_export_symbols=$export_symbols export_symbols= always_export_symbols=yes } fi ;; esac # Prepare the list of exported symbols if test -z "$export_symbols"; then if test yes = "$always_export_symbols" || test -n "$export_symbols_regex"; then func_verbose "generating symbol list for '$libname.la'" export_symbols=$output_objdir/$libname.exp $opt_dry_run || $RM $export_symbols cmds=$export_symbols_cmds save_ifs=$IFS; IFS='~' for cmd1 in $cmds; do IFS=$save_ifs # Take the normal branch if the nm_file_list_spec branch # doesn't work or if tool conversion is not needed. case $nm_file_list_spec~$to_tool_file_cmd in *~func_convert_file_noop | *~func_convert_file_msys_to_w32 | ~*) try_normal_branch=yes eval cmd=\"$cmd1\" func_len " $cmd" len=$func_len_result ;; *) try_normal_branch=no ;; esac if test yes = "$try_normal_branch" \ && { test "$len" -lt "$max_cmd_len" \ || test "$max_cmd_len" -le -1; } then func_show_eval "$cmd" 'exit $?' skipped_export=false elif test -n "$nm_file_list_spec"; then func_basename "$output" output_la=$func_basename_result save_libobjs=$libobjs save_output=$output output=$output_objdir/$output_la.nm func_to_tool_file "$output" libobjs=$nm_file_list_spec$func_to_tool_file_result func_append delfiles " $output" func_verbose "creating $NM input file list: $output" for obj in $save_libobjs; do func_to_tool_file "$obj" $ECHO "$func_to_tool_file_result" done > "$output" eval cmd=\"$cmd1\" func_show_eval "$cmd" 'exit $?' output=$save_output libobjs=$save_libobjs skipped_export=false else # The command line is too long to execute in one step. func_verbose "using reloadable object file for export list..." skipped_export=: # Break out early, otherwise skipped_export may be # set to false by a later but shorter cmd. break fi done IFS=$save_ifs if test -n "$export_symbols_regex" && test : != "$skipped_export"; then func_show_eval '$EGREP -e "$export_symbols_regex" "$export_symbols" > "${export_symbols}T"' func_show_eval '$MV "${export_symbols}T" "$export_symbols"' fi fi fi if test -n "$export_symbols" && test -n "$include_expsyms"; then tmp_export_symbols=$export_symbols test -n "$orig_export_symbols" && tmp_export_symbols=$orig_export_symbols $opt_dry_run || eval '$ECHO "$include_expsyms" | $SP2NL >> "$tmp_export_symbols"' fi if test : != "$skipped_export" && test -n "$orig_export_symbols"; then # The given exports_symbols file has to be filtered, so filter it. func_verbose "filter symbol list for '$libname.la' to tag DATA exports" # FIXME: $output_objdir/$libname.filter potentially contains lots of # 's' commands, which not all seds can handle. GNU sed should be fine # though. Also, the filter scales superlinearly with the number of # global variables. join(1) would be nice here, but unfortunately # isn't a blessed tool. $opt_dry_run || $SED -e '/[ ,]DATA/!d;s,\(.*\)\([ \,].*\),s|^\1$|\1\2|,' < $export_symbols > $output_objdir/$libname.filter func_append delfiles " $export_symbols $output_objdir/$libname.filter" export_symbols=$output_objdir/$libname.def $opt_dry_run || $SED -f $output_objdir/$libname.filter < $orig_export_symbols > $export_symbols fi tmp_deplibs= for test_deplib in $deplibs; do case " $convenience " in *" $test_deplib "*) ;; *) func_append tmp_deplibs " $test_deplib" ;; esac done deplibs=$tmp_deplibs if test -n "$convenience"; then if test -n "$whole_archive_flag_spec" && test yes = "$compiler_needs_object" && test -z "$libobjs"; then # extract the archives, so we have objects to list. # TODO: could optimize this to just extract one archive. whole_archive_flag_spec= fi if test -n "$whole_archive_flag_spec"; then save_libobjs=$libobjs eval libobjs=\"\$libobjs $whole_archive_flag_spec\" test "X$libobjs" = "X " && libobjs= else gentop=$output_objdir/${outputname}x func_append generated " $gentop" func_extract_archives $gentop $convenience func_append libobjs " $func_extract_archives_result" test "X$libobjs" = "X " && libobjs= fi fi if test yes = "$thread_safe" && test -n "$thread_safe_flag_spec"; then eval flag=\"$thread_safe_flag_spec\" func_append linker_flags " $flag" fi # Make a backup of the uninstalled library when relinking if test relink = "$opt_mode"; then $opt_dry_run || eval '(cd $output_objdir && $RM ${realname}U && $MV $realname ${realname}U)' || exit $? fi # Do each of the archive commands. if test yes = "$module" && test -n "$module_cmds"; then if test -n "$export_symbols" && test -n "$module_expsym_cmds"; then eval test_cmds=\"$module_expsym_cmds\" cmds=$module_expsym_cmds else eval test_cmds=\"$module_cmds\" cmds=$module_cmds fi else if test -n "$export_symbols" && test -n "$archive_expsym_cmds"; then eval test_cmds=\"$archive_expsym_cmds\" cmds=$archive_expsym_cmds else eval test_cmds=\"$archive_cmds\" cmds=$archive_cmds fi fi if test : != "$skipped_export" && func_len " $test_cmds" && len=$func_len_result && test "$len" -lt "$max_cmd_len" || test "$max_cmd_len" -le -1; then : else # The command line is too long to link in one step, link piecewise # or, if using GNU ld and skipped_export is not :, use a linker # script. # Save the value of $output and $libobjs because we want to # use them later. If we have whole_archive_flag_spec, we # want to use save_libobjs as it was before # whole_archive_flag_spec was expanded, because we can't # assume the linker understands whole_archive_flag_spec. # This may have to be revisited, in case too many # convenience libraries get linked in and end up exceeding # the spec. if test -z "$convenience" || test -z "$whole_archive_flag_spec"; then save_libobjs=$libobjs fi save_output=$output func_basename "$output" output_la=$func_basename_result # Clear the reloadable object creation command queue and # initialize k to one. test_cmds= concat_cmds= objlist= last_robj= k=1 if test -n "$save_libobjs" && test : != "$skipped_export" && test yes = "$with_gnu_ld"; then output=$output_objdir/$output_la.lnkscript func_verbose "creating GNU ld script: $output" echo 'INPUT (' > $output for obj in $save_libobjs do func_to_tool_file "$obj" $ECHO "$func_to_tool_file_result" >> $output done echo ')' >> $output func_append delfiles " $output" func_to_tool_file "$output" output=$func_to_tool_file_result elif test -n "$save_libobjs" && test : != "$skipped_export" && test -n "$file_list_spec"; then output=$output_objdir/$output_la.lnk func_verbose "creating linker input file list: $output" : > $output set x $save_libobjs shift firstobj= if test yes = "$compiler_needs_object"; then firstobj="$1 " shift fi for obj do func_to_tool_file "$obj" $ECHO "$func_to_tool_file_result" >> $output done func_append delfiles " $output" func_to_tool_file "$output" output=$firstobj\"$file_list_spec$func_to_tool_file_result\" else if test -n "$save_libobjs"; then func_verbose "creating reloadable object files..." output=$output_objdir/$output_la-$k.$objext eval test_cmds=\"$reload_cmds\" func_len " $test_cmds" len0=$func_len_result len=$len0 # Loop over the list of objects to be linked. for obj in $save_libobjs do func_len " $obj" func_arith $len + $func_len_result len=$func_arith_result if test -z "$objlist" || test "$len" -lt "$max_cmd_len"; then func_append objlist " $obj" else # The command $test_cmds is almost too long, add a # command to the queue. if test 1 -eq "$k"; then # The first file doesn't have a previous command to add. reload_objs=$objlist eval concat_cmds=\"$reload_cmds\" else # All subsequent reloadable object files will link in # the last one created. reload_objs="$objlist $last_robj" eval concat_cmds=\"\$concat_cmds~$reload_cmds~\$RM $last_robj\" fi last_robj=$output_objdir/$output_la-$k.$objext func_arith $k + 1 k=$func_arith_result output=$output_objdir/$output_la-$k.$objext objlist=" $obj" func_len " $last_robj" func_arith $len0 + $func_len_result len=$func_arith_result fi done # Handle the remaining objects by creating one last # reloadable object file. All subsequent reloadable object # files will link in the last one created. test -z "$concat_cmds" || concat_cmds=$concat_cmds~ reload_objs="$objlist $last_robj" eval concat_cmds=\"\$concat_cmds$reload_cmds\" if test -n "$last_robj"; then eval concat_cmds=\"\$concat_cmds~\$RM $last_robj\" fi func_append delfiles " $output" else output= fi ${skipped_export-false} && { func_verbose "generating symbol list for '$libname.la'" export_symbols=$output_objdir/$libname.exp $opt_dry_run || $RM $export_symbols libobjs=$output # Append the command to create the export file. test -z "$concat_cmds" || concat_cmds=$concat_cmds~ eval concat_cmds=\"\$concat_cmds$export_symbols_cmds\" if test -n "$last_robj"; then eval concat_cmds=\"\$concat_cmds~\$RM $last_robj\" fi } test -n "$save_libobjs" && func_verbose "creating a temporary reloadable object file: $output" # Loop through the commands generated above and execute them. save_ifs=$IFS; IFS='~' for cmd in $concat_cmds; do IFS=$save_ifs $opt_quiet || { func_quote_for_expand "$cmd" eval "func_echo $func_quote_for_expand_result" } $opt_dry_run || eval "$cmd" || { lt_exit=$? # Restore the uninstalled library and exit if test relink = "$opt_mode"; then ( cd "$output_objdir" && \ $RM "${realname}T" && \ $MV "${realname}U" "$realname" ) fi exit $lt_exit } done IFS=$save_ifs if test -n "$export_symbols_regex" && ${skipped_export-false}; then func_show_eval '$EGREP -e "$export_symbols_regex" "$export_symbols" > "${export_symbols}T"' func_show_eval '$MV "${export_symbols}T" "$export_symbols"' fi fi ${skipped_export-false} && { if test -n "$export_symbols" && test -n "$include_expsyms"; then tmp_export_symbols=$export_symbols test -n "$orig_export_symbols" && tmp_export_symbols=$orig_export_symbols $opt_dry_run || eval '$ECHO "$include_expsyms" | $SP2NL >> "$tmp_export_symbols"' fi if test -n "$orig_export_symbols"; then # The given exports_symbols file has to be filtered, so filter it. func_verbose "filter symbol list for '$libname.la' to tag DATA exports" # FIXME: $output_objdir/$libname.filter potentially contains lots of # 's' commands, which not all seds can handle. GNU sed should be fine # though. Also, the filter scales superlinearly with the number of # global variables. join(1) would be nice here, but unfortunately # isn't a blessed tool. $opt_dry_run || $SED -e '/[ ,]DATA/!d;s,\(.*\)\([ \,].*\),s|^\1$|\1\2|,' < $export_symbols > $output_objdir/$libname.filter func_append delfiles " $export_symbols $output_objdir/$libname.filter" export_symbols=$output_objdir/$libname.def $opt_dry_run || $SED -f $output_objdir/$libname.filter < $orig_export_symbols > $export_symbols fi } libobjs=$output # Restore the value of output. output=$save_output if test -n "$convenience" && test -n "$whole_archive_flag_spec"; then eval libobjs=\"\$libobjs $whole_archive_flag_spec\" test "X$libobjs" = "X " && libobjs= fi # Expand the library linking commands again to reset the # value of $libobjs for piecewise linking. # Do each of the archive commands. if test yes = "$module" && test -n "$module_cmds"; then if test -n "$export_symbols" && test -n "$module_expsym_cmds"; then cmds=$module_expsym_cmds else cmds=$module_cmds fi else if test -n "$export_symbols" && test -n "$archive_expsym_cmds"; then cmds=$archive_expsym_cmds else cmds=$archive_cmds fi fi fi if test -n "$delfiles"; then # Append the command to remove temporary files to $cmds. eval cmds=\"\$cmds~\$RM $delfiles\" fi # Add any objects from preloaded convenience libraries if test -n "$dlprefiles"; then gentop=$output_objdir/${outputname}x func_append generated " $gentop" func_extract_archives $gentop $dlprefiles func_append libobjs " $func_extract_archives_result" test "X$libobjs" = "X " && libobjs= fi save_ifs=$IFS; IFS='~' for cmd in $cmds; do IFS=$sp$nl eval cmd=\"$cmd\" IFS=$save_ifs $opt_quiet || { func_quote_for_expand "$cmd" eval "func_echo $func_quote_for_expand_result" } $opt_dry_run || eval "$cmd" || { lt_exit=$? # Restore the uninstalled library and exit if test relink = "$opt_mode"; then ( cd "$output_objdir" && \ $RM "${realname}T" && \ $MV "${realname}U" "$realname" ) fi exit $lt_exit } done IFS=$save_ifs # Restore the uninstalled library and exit if test relink = "$opt_mode"; then $opt_dry_run || eval '(cd $output_objdir && $RM ${realname}T && $MV $realname ${realname}T && $MV ${realname}U $realname)' || exit $? if test -n "$convenience"; then if test -z "$whole_archive_flag_spec"; then func_show_eval '${RM}r "$gentop"' fi fi exit $EXIT_SUCCESS fi # Create links to the real library. for linkname in $linknames; do if test "$realname" != "$linkname"; then func_show_eval '(cd "$output_objdir" && $RM "$linkname" && $LN_S "$realname" "$linkname")' 'exit $?' fi done # If -module or -export-dynamic was specified, set the dlname. if test yes = "$module" || test yes = "$export_dynamic"; then # On all known operating systems, these are identical. dlname=$soname fi fi ;; obj) if test -n "$dlfiles$dlprefiles" || test no != "$dlself"; then func_warning "'-dlopen' is ignored for objects" fi case " $deplibs" in *\ -l* | *\ -L*) func_warning "'-l' and '-L' are ignored for objects" ;; esac test -n "$rpath" && \ func_warning "'-rpath' is ignored for objects" test -n "$xrpath" && \ func_warning "'-R' is ignored for objects" test -n "$vinfo" && \ func_warning "'-version-info' is ignored for objects" test -n "$release" && \ func_warning "'-release' is ignored for objects" case $output in *.lo) test -n "$objs$old_deplibs" && \ func_fatal_error "cannot build library object '$output' from non-libtool objects" libobj=$output func_lo2o "$libobj" obj=$func_lo2o_result ;; *) libobj= obj=$output ;; esac # Delete the old objects. $opt_dry_run || $RM $obj $libobj # Objects from convenience libraries. This assumes # single-version convenience libraries. Whenever we create # different ones for PIC/non-PIC, this we'll have to duplicate # the extraction. reload_conv_objs= gentop= # if reload_cmds runs $LD directly, get rid of -Wl from # whole_archive_flag_spec and hope we can get by with turning comma # into space. case $reload_cmds in *\$LD[\ \$]*) wl= ;; esac if test -n "$convenience"; then if test -n "$whole_archive_flag_spec"; then eval tmp_whole_archive_flags=\"$whole_archive_flag_spec\" test -n "$wl" || tmp_whole_archive_flags=`$ECHO "$tmp_whole_archive_flags" | $SED 's|,| |g'` reload_conv_objs=$reload_objs\ $tmp_whole_archive_flags else gentop=$output_objdir/${obj}x func_append generated " $gentop" func_extract_archives $gentop $convenience reload_conv_objs="$reload_objs $func_extract_archives_result" fi fi # If we're not building shared, we need to use non_pic_objs test yes = "$build_libtool_libs" || libobjs=$non_pic_objects # Create the old-style object. reload_objs=$objs$old_deplibs' '`$ECHO "$libobjs" | $SP2NL | $SED "/\.$libext$/d; /\.lib$/d; $lo2o" | $NL2SP`' '$reload_conv_objs output=$obj func_execute_cmds "$reload_cmds" 'exit $?' # Exit if we aren't doing a library object file. if test -z "$libobj"; then if test -n "$gentop"; then func_show_eval '${RM}r "$gentop"' fi exit $EXIT_SUCCESS fi test yes = "$build_libtool_libs" || { if test -n "$gentop"; then func_show_eval '${RM}r "$gentop"' fi # Create an invalid libtool object if no PIC, so that we don't # accidentally link it into a program. # $show "echo timestamp > $libobj" # $opt_dry_run || eval "echo timestamp > $libobj" || exit $? exit $EXIT_SUCCESS } if test -n "$pic_flag" || test default != "$pic_mode"; then # Only do commands if we really have different PIC objects. reload_objs="$libobjs $reload_conv_objs" output=$libobj func_execute_cmds "$reload_cmds" 'exit $?' fi if test -n "$gentop"; then func_show_eval '${RM}r "$gentop"' fi exit $EXIT_SUCCESS ;; prog) case $host in *cygwin*) func_stripname '' '.exe' "$output" output=$func_stripname_result.exe;; esac test -n "$vinfo" && \ func_warning "'-version-info' is ignored for programs" test -n "$release" && \ func_warning "'-release' is ignored for programs" $preload \ && test unknown,unknown,unknown = "$dlopen_support,$dlopen_self,$dlopen_self_static" \ && func_warning "'LT_INIT([dlopen])' not used. Assuming no dlopen support." case $host in *-*-rhapsody* | *-*-darwin1.[012]) # On Rhapsody replace the C library is the System framework compile_deplibs=`$ECHO " $compile_deplibs" | $SED 's/ -lc / System.ltframework /'` finalize_deplibs=`$ECHO " $finalize_deplibs" | $SED 's/ -lc / System.ltframework /'` ;; esac case $host in *-*-darwin*) # Don't allow lazy linking, it breaks C++ global constructors # But is supposedly fixed on 10.4 or later (yay!). if test CXX = "$tagname"; then case ${MACOSX_DEPLOYMENT_TARGET-10.0} in 10.[0123]) func_append compile_command " $wl-bind_at_load" func_append finalize_command " $wl-bind_at_load" ;; esac fi # Time to change all our "foo.ltframework" stuff back to "-framework foo" compile_deplibs=`$ECHO " $compile_deplibs" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` finalize_deplibs=`$ECHO " $finalize_deplibs" | $SED 's% \([^ $]*\).ltframework% -framework \1%g'` ;; esac # move library search paths that coincide with paths to not yet # installed libraries to the beginning of the library search list new_libs= for path in $notinst_path; do case " $new_libs " in *" -L$path/$objdir "*) ;; *) case " $compile_deplibs " in *" -L$path/$objdir "*) func_append new_libs " -L$path/$objdir" ;; esac ;; esac done for deplib in $compile_deplibs; do case $deplib in -L*) case " $new_libs " in *" $deplib "*) ;; *) func_append new_libs " $deplib" ;; esac ;; *) func_append new_libs " $deplib" ;; esac done compile_deplibs=$new_libs func_append compile_command " $compile_deplibs" func_append finalize_command " $finalize_deplibs" if test -n "$rpath$xrpath"; then # If the user specified any rpath flags, then add them. for libdir in $rpath $xrpath; do # This is the magic to use -rpath. case "$finalize_rpath " in *" $libdir "*) ;; *) func_append finalize_rpath " $libdir" ;; esac done fi # Now hardcode the library paths rpath= hardcode_libdirs= for libdir in $compile_rpath $finalize_rpath; do if test -n "$hardcode_libdir_flag_spec"; then if test -n "$hardcode_libdir_separator"; then if test -z "$hardcode_libdirs"; then hardcode_libdirs=$libdir else # Just accumulate the unique libdirs. case $hardcode_libdir_separator$hardcode_libdirs$hardcode_libdir_separator in *"$hardcode_libdir_separator$libdir$hardcode_libdir_separator"*) ;; *) func_append hardcode_libdirs "$hardcode_libdir_separator$libdir" ;; esac fi else eval flag=\"$hardcode_libdir_flag_spec\" func_append rpath " $flag" fi elif test -n "$runpath_var"; then case "$perm_rpath " in *" $libdir "*) ;; *) func_append perm_rpath " $libdir" ;; esac fi case $host in *-*-cygwin* | *-*-mingw* | *-*-pw32* | *-*-os2* | *-cegcc*) testbindir=`$ECHO "$libdir" | $SED -e 's*/lib$*/bin*'` case :$dllsearchpath: in *":$libdir:"*) ;; ::) dllsearchpath=$libdir;; *) func_append dllsearchpath ":$libdir";; esac case :$dllsearchpath: in *":$testbindir:"*) ;; ::) dllsearchpath=$testbindir;; *) func_append dllsearchpath ":$testbindir";; esac ;; esac done # Substitute the hardcoded libdirs into the rpath. if test -n "$hardcode_libdir_separator" && test -n "$hardcode_libdirs"; then libdir=$hardcode_libdirs eval rpath=\" $hardcode_libdir_flag_spec\" fi compile_rpath=$rpath rpath= hardcode_libdirs= for libdir in $finalize_rpath; do if test -n "$hardcode_libdir_flag_spec"; then if test -n "$hardcode_libdir_separator"; then if test -z "$hardcode_libdirs"; then hardcode_libdirs=$libdir else # Just accumulate the unique libdirs. case $hardcode_libdir_separator$hardcode_libdirs$hardcode_libdir_separator in *"$hardcode_libdir_separator$libdir$hardcode_libdir_separator"*) ;; *) func_append hardcode_libdirs "$hardcode_libdir_separator$libdir" ;; esac fi else eval flag=\"$hardcode_libdir_flag_spec\" func_append rpath " $flag" fi elif test -n "$runpath_var"; then case "$finalize_perm_rpath " in *" $libdir "*) ;; *) func_append finalize_perm_rpath " $libdir" ;; esac fi done # Substitute the hardcoded libdirs into the rpath. if test -n "$hardcode_libdir_separator" && test -n "$hardcode_libdirs"; then libdir=$hardcode_libdirs eval rpath=\" $hardcode_libdir_flag_spec\" fi finalize_rpath=$rpath if test -n "$libobjs" && test yes = "$build_old_libs"; then # Transform all the library objects into standard objects. compile_command=`$ECHO "$compile_command" | $SP2NL | $SED "$lo2o" | $NL2SP` finalize_command=`$ECHO "$finalize_command" | $SP2NL | $SED "$lo2o" | $NL2SP` fi func_generate_dlsyms "$outputname" "@PROGRAM@" false # template prelinking step if test -n "$prelink_cmds"; then func_execute_cmds "$prelink_cmds" 'exit $?' fi wrappers_required=: case $host in *cegcc* | *mingw32ce*) # Disable wrappers for cegcc and mingw32ce hosts, we are cross compiling anyway. wrappers_required=false ;; *cygwin* | *mingw* ) test yes = "$build_libtool_libs" || wrappers_required=false ;; *) if test no = "$need_relink" || test yes != "$build_libtool_libs"; then wrappers_required=false fi ;; esac $wrappers_required || { # Replace the output file specification. compile_command=`$ECHO "$compile_command" | $SED 's%@OUTPUT@%'"$output"'%g'` link_command=$compile_command$compile_rpath # We have no uninstalled library dependencies, so finalize right now. exit_status=0 func_show_eval "$link_command" 'exit_status=$?' if test -n "$postlink_cmds"; then func_to_tool_file "$output" postlink_cmds=`func_echo_all "$postlink_cmds" | $SED -e 's%@OUTPUT@%'"$output"'%g' -e 's%@TOOL_OUTPUT@%'"$func_to_tool_file_result"'%g'` func_execute_cmds "$postlink_cmds" 'exit $?' fi # Delete the generated files. if test -f "$output_objdir/${outputname}S.$objext"; then func_show_eval '$RM "$output_objdir/${outputname}S.$objext"' fi exit $exit_status } if test -n "$compile_shlibpath$finalize_shlibpath"; then compile_command="$shlibpath_var=\"$compile_shlibpath$finalize_shlibpath\$$shlibpath_var\" $compile_command" fi if test -n "$finalize_shlibpath"; then finalize_command="$shlibpath_var=\"$finalize_shlibpath\$$shlibpath_var\" $finalize_command" fi compile_var= finalize_var= if test -n "$runpath_var"; then if test -n "$perm_rpath"; then # We should set the runpath_var. rpath= for dir in $perm_rpath; do func_append rpath "$dir:" done compile_var="$runpath_var=\"$rpath\$$runpath_var\" " fi if test -n "$finalize_perm_rpath"; then # We should set the runpath_var. rpath= for dir in $finalize_perm_rpath; do func_append rpath "$dir:" done finalize_var="$runpath_var=\"$rpath\$$runpath_var\" " fi fi if test yes = "$no_install"; then # We don't need to create a wrapper script. link_command=$compile_var$compile_command$compile_rpath # Replace the output file specification. link_command=`$ECHO "$link_command" | $SED 's%@OUTPUT@%'"$output"'%g'` # Delete the old output file. $opt_dry_run || $RM $output # Link the executable and exit func_show_eval "$link_command" 'exit $?' if test -n "$postlink_cmds"; then func_to_tool_file "$output" postlink_cmds=`func_echo_all "$postlink_cmds" | $SED -e 's%@OUTPUT@%'"$output"'%g' -e 's%@TOOL_OUTPUT@%'"$func_to_tool_file_result"'%g'` func_execute_cmds "$postlink_cmds" 'exit $?' fi exit $EXIT_SUCCESS fi case $hardcode_action,$fast_install in relink,*) # Fast installation is not supported link_command=$compile_var$compile_command$compile_rpath relink_command=$finalize_var$finalize_command$finalize_rpath func_warning "this platform does not like uninstalled shared libraries" func_warning "'$output' will be relinked during installation" ;; *,yes) link_command=$finalize_var$compile_command$finalize_rpath relink_command=`$ECHO "$compile_var$compile_command$compile_rpath" | $SED 's%@OUTPUT@%\$progdir/\$file%g'` ;; *,no) link_command=$compile_var$compile_command$compile_rpath relink_command=$finalize_var$finalize_command$finalize_rpath ;; *,needless) link_command=$finalize_var$compile_command$finalize_rpath relink_command= ;; esac # Replace the output file specification. link_command=`$ECHO "$link_command" | $SED 's%@OUTPUT@%'"$output_objdir/$outputname"'%g'` # Delete the old output files. $opt_dry_run || $RM $output $output_objdir/$outputname $output_objdir/lt-$outputname func_show_eval "$link_command" 'exit $?' if test -n "$postlink_cmds"; then func_to_tool_file "$output_objdir/$outputname" postlink_cmds=`func_echo_all "$postlink_cmds" | $SED -e 's%@OUTPUT@%'"$output_objdir/$outputname"'%g' -e 's%@TOOL_OUTPUT@%'"$func_to_tool_file_result"'%g'` func_execute_cmds "$postlink_cmds" 'exit $?' fi # Now create the wrapper script. func_verbose "creating $output" # Quote the relink command for shipping. if test -n "$relink_command"; then # Preserve any variables that may affect compiler behavior for var in $variables_saved_for_relink; do if eval test -z \"\${$var+set}\"; then relink_command="{ test -z \"\${$var+set}\" || $lt_unset $var || { $var=; export $var; }; }; $relink_command" elif eval var_value=\$$var; test -z "$var_value"; then relink_command="$var=; export $var; $relink_command" else func_quote_for_eval "$var_value" relink_command="$var=$func_quote_for_eval_result; export $var; $relink_command" fi done relink_command="(cd `pwd`; $relink_command)" relink_command=`$ECHO "$relink_command" | $SED "$sed_quote_subst"` fi # Only actually do things if not in dry run mode. $opt_dry_run || { # win32 will think the script is a binary if it has # a .exe suffix, so we strip it off here. case $output in *.exe) func_stripname '' '.exe' "$output" output=$func_stripname_result ;; esac # test for cygwin because mv fails w/o .exe extensions case $host in *cygwin*) exeext=.exe func_stripname '' '.exe' "$outputname" outputname=$func_stripname_result ;; *) exeext= ;; esac case $host in *cygwin* | *mingw* ) func_dirname_and_basename "$output" "" "." output_name=$func_basename_result output_path=$func_dirname_result cwrappersource=$output_path/$objdir/lt-$output_name.c cwrapper=$output_path/$output_name.exe $RM $cwrappersource $cwrapper trap "$RM $cwrappersource $cwrapper; exit $EXIT_FAILURE" 1 2 15 func_emit_cwrapperexe_src > $cwrappersource # The wrapper executable is built using the $host compiler, # because it contains $host paths and files. If cross- # compiling, it, like the target executable, must be # executed on the $host or under an emulation environment. $opt_dry_run || { $LTCC $LTCFLAGS -o $cwrapper $cwrappersource $STRIP $cwrapper } # Now, create the wrapper script for func_source use: func_ltwrapper_scriptname $cwrapper $RM $func_ltwrapper_scriptname_result trap "$RM $func_ltwrapper_scriptname_result; exit $EXIT_FAILURE" 1 2 15 $opt_dry_run || { # note: this script will not be executed, so do not chmod. if test "x$build" = "x$host"; then $cwrapper --lt-dump-script > $func_ltwrapper_scriptname_result else func_emit_wrapper no > $func_ltwrapper_scriptname_result fi } ;; * ) $RM $output trap "$RM $output; exit $EXIT_FAILURE" 1 2 15 func_emit_wrapper no > $output chmod +x $output ;; esac } exit $EXIT_SUCCESS ;; esac # See if we need to build an old-fashioned archive. for oldlib in $oldlibs; do case $build_libtool_libs in convenience) oldobjs="$libobjs_save $symfileobj" addlibs=$convenience build_libtool_libs=no ;; module) oldobjs=$libobjs_save addlibs=$old_convenience build_libtool_libs=no ;; *) oldobjs="$old_deplibs $non_pic_objects" $preload && test -f "$symfileobj" \ && func_append oldobjs " $symfileobj" addlibs=$old_convenience ;; esac if test -n "$addlibs"; then gentop=$output_objdir/${outputname}x func_append generated " $gentop" func_extract_archives $gentop $addlibs func_append oldobjs " $func_extract_archives_result" fi # Do each command in the archive commands. if test -n "$old_archive_from_new_cmds" && test yes = "$build_libtool_libs"; then cmds=$old_archive_from_new_cmds else # Add any objects from preloaded convenience libraries if test -n "$dlprefiles"; then gentop=$output_objdir/${outputname}x func_append generated " $gentop" func_extract_archives $gentop $dlprefiles func_append oldobjs " $func_extract_archives_result" fi # POSIX demands no paths to be encoded in archives. We have # to avoid creating archives with duplicate basenames if we # might have to extract them afterwards, e.g., when creating a # static archive out of a convenience library, or when linking # the entirety of a libtool archive into another (currently # not supported by libtool). if (for obj in $oldobjs do func_basename "$obj" $ECHO "$func_basename_result" done | sort | sort -uc >/dev/null 2>&1); then : else echo "copying selected object files to avoid basename conflicts..." gentop=$output_objdir/${outputname}x func_append generated " $gentop" func_mkdir_p "$gentop" save_oldobjs=$oldobjs oldobjs= counter=1 for obj in $save_oldobjs do func_basename "$obj" objbase=$func_basename_result case " $oldobjs " in " ") oldobjs=$obj ;; *[\ /]"$objbase "*) while :; do # Make sure we don't pick an alternate name that also # overlaps. newobj=lt$counter-$objbase func_arith $counter + 1 counter=$func_arith_result case " $oldobjs " in *[\ /]"$newobj "*) ;; *) if test ! -f "$gentop/$newobj"; then break; fi ;; esac done func_show_eval "ln $obj $gentop/$newobj || cp $obj $gentop/$newobj" func_append oldobjs " $gentop/$newobj" ;; *) func_append oldobjs " $obj" ;; esac done fi func_to_tool_file "$oldlib" func_convert_file_msys_to_w32 tool_oldlib=$func_to_tool_file_result eval cmds=\"$old_archive_cmds\" func_len " $cmds" len=$func_len_result if test "$len" -lt "$max_cmd_len" || test "$max_cmd_len" -le -1; then cmds=$old_archive_cmds elif test -n "$archiver_list_spec"; then func_verbose "using command file archive linking..." for obj in $oldobjs do func_to_tool_file "$obj" $ECHO "$func_to_tool_file_result" done > $output_objdir/$libname.libcmd func_to_tool_file "$output_objdir/$libname.libcmd" oldobjs=" $archiver_list_spec$func_to_tool_file_result" cmds=$old_archive_cmds else # the command line is too long to link in one step, link in parts func_verbose "using piecewise archive linking..." save_RANLIB=$RANLIB RANLIB=: objlist= concat_cmds= save_oldobjs=$oldobjs oldobjs= # Is there a better way of finding the last object in the list? for obj in $save_oldobjs do last_oldobj=$obj done eval test_cmds=\"$old_archive_cmds\" func_len " $test_cmds" len0=$func_len_result len=$len0 for obj in $save_oldobjs do func_len " $obj" func_arith $len + $func_len_result len=$func_arith_result func_append objlist " $obj" if test "$len" -lt "$max_cmd_len"; then : else # the above command should be used before it gets too long oldobjs=$objlist if test "$obj" = "$last_oldobj"; then RANLIB=$save_RANLIB fi test -z "$concat_cmds" || concat_cmds=$concat_cmds~ eval concat_cmds=\"\$concat_cmds$old_archive_cmds\" objlist= len=$len0 fi done RANLIB=$save_RANLIB oldobjs=$objlist if test -z "$oldobjs"; then eval cmds=\"\$concat_cmds\" else eval cmds=\"\$concat_cmds~\$old_archive_cmds\" fi fi fi func_execute_cmds "$cmds" 'exit $?' done test -n "$generated" && \ func_show_eval "${RM}r$generated" # Now create the libtool archive. case $output in *.la) old_library= test yes = "$build_old_libs" && old_library=$libname.$libext func_verbose "creating $output" # Preserve any variables that may affect compiler behavior for var in $variables_saved_for_relink; do if eval test -z \"\${$var+set}\"; then relink_command="{ test -z \"\${$var+set}\" || $lt_unset $var || { $var=; export $var; }; }; $relink_command" elif eval var_value=\$$var; test -z "$var_value"; then relink_command="$var=; export $var; $relink_command" else func_quote_for_eval "$var_value" relink_command="$var=$func_quote_for_eval_result; export $var; $relink_command" fi done # Quote the link command for shipping. relink_command="(cd `pwd`; $SHELL \"$progpath\" $preserve_args --mode=relink $libtool_args @inst_prefix_dir@)" relink_command=`$ECHO "$relink_command" | $SED "$sed_quote_subst"` if test yes = "$hardcode_automatic"; then relink_command= fi # Only create the output if not a dry run. $opt_dry_run || { for installed in no yes; do if test yes = "$installed"; then if test -z "$install_libdir"; then break fi output=$output_objdir/${outputname}i # Replace all uninstalled libtool libraries with the installed ones newdependency_libs= for deplib in $dependency_libs; do case $deplib in *.la) func_basename "$deplib" name=$func_basename_result func_resolve_sysroot "$deplib" eval libdir=`$SED -n -e 's/^libdir=\(.*\)$/\1/p' $func_resolve_sysroot_result` test -z "$libdir" && \ func_fatal_error "'$deplib' is not a valid libtool archive" func_append newdependency_libs " ${lt_sysroot:+=}$libdir/$name" ;; -L*) func_stripname -L '' "$deplib" func_replace_sysroot "$func_stripname_result" func_append newdependency_libs " -L$func_replace_sysroot_result" ;; -R*) func_stripname -R '' "$deplib" func_replace_sysroot "$func_stripname_result" func_append newdependency_libs " -R$func_replace_sysroot_result" ;; *) func_append newdependency_libs " $deplib" ;; esac done dependency_libs=$newdependency_libs newdlfiles= for lib in $dlfiles; do case $lib in *.la) func_basename "$lib" name=$func_basename_result eval libdir=`$SED -n -e 's/^libdir=\(.*\)$/\1/p' $lib` test -z "$libdir" && \ func_fatal_error "'$lib' is not a valid libtool archive" func_append newdlfiles " ${lt_sysroot:+=}$libdir/$name" ;; *) func_append newdlfiles " $lib" ;; esac done dlfiles=$newdlfiles newdlprefiles= for lib in $dlprefiles; do case $lib in *.la) # Only pass preopened files to the pseudo-archive (for # eventual linking with the app. that links it) if we # didn't already link the preopened objects directly into # the library: func_basename "$lib" name=$func_basename_result eval libdir=`$SED -n -e 's/^libdir=\(.*\)$/\1/p' $lib` test -z "$libdir" && \ func_fatal_error "'$lib' is not a valid libtool archive" func_append newdlprefiles " ${lt_sysroot:+=}$libdir/$name" ;; esac done dlprefiles=$newdlprefiles else newdlfiles= for lib in $dlfiles; do case $lib in [\\/]* | [A-Za-z]:[\\/]*) abs=$lib ;; *) abs=`pwd`"/$lib" ;; esac func_append newdlfiles " $abs" done dlfiles=$newdlfiles newdlprefiles= for lib in $dlprefiles; do case $lib in [\\/]* | [A-Za-z]:[\\/]*) abs=$lib ;; *) abs=`pwd`"/$lib" ;; esac func_append newdlprefiles " $abs" done dlprefiles=$newdlprefiles fi $RM $output # place dlname in correct position for cygwin # In fact, it would be nice if we could use this code for all target # systems that can't hard-code library paths into their executables # and that have no shared library path variable independent of PATH, # but it turns out we can't easily determine that from inspecting # libtool variables, so we have to hard-code the OSs to which it # applies here; at the moment, that means platforms that use the PE # object format with DLL files. See the long comment at the top of # tests/bindir.at for full details. tdlname=$dlname case $host,$output,$installed,$module,$dlname in *cygwin*,*lai,yes,no,*.dll | *mingw*,*lai,yes,no,*.dll | *cegcc*,*lai,yes,no,*.dll) # If a -bindir argument was supplied, place the dll there. if test -n "$bindir"; then func_relative_path "$install_libdir" "$bindir" tdlname=$func_relative_path_result/$dlname else # Otherwise fall back on heuristic. tdlname=../bin/$dlname fi ;; esac $ECHO > $output "\ # $outputname - a libtool library file # Generated by $PROGRAM (GNU $PACKAGE) $VERSION # # Please DO NOT delete this file! # It is necessary for linking the library. # The name that we can dlopen(3). dlname='$tdlname' # Names of this library. library_names='$library_names' # The name of the static archive. old_library='$old_library' # Linker flags that cannot go in dependency_libs. inherited_linker_flags='$new_inherited_linker_flags' # Libraries that this one depends upon. dependency_libs='$dependency_libs' # Names of additional weak libraries provided by this library weak_library_names='$weak_libs' # Version information for $libname. current=$current age=$age revision=$revision # Is this an already installed library? installed=$installed # Should we warn about portability when linking against -modules? shouldnotlink=$module # Files to dlopen/dlpreopen dlopen='$dlfiles' dlpreopen='$dlprefiles' # Directory that this library needs to be installed in: libdir='$install_libdir'" if test no,yes = "$installed,$need_relink"; then $ECHO >> $output "\ relink_command=\"$relink_command\"" fi done } # Do a symbolic link so that the libtool archive can be found in # LD_LIBRARY_PATH before the program is installed. func_show_eval '( cd "$output_objdir" && $RM "$outputname" && $LN_S "../$outputname" "$outputname" )' 'exit $?' ;; esac exit $EXIT_SUCCESS } if test link = "$opt_mode" || test relink = "$opt_mode"; then func_mode_link ${1+"$@"} fi # func_mode_uninstall arg... func_mode_uninstall () { $debug_cmd RM=$nonopt files= rmforce=false exit_status=0 # This variable tells wrapper scripts just to set variables rather # than running their programs. libtool_install_magic=$magic for arg do case $arg in -f) func_append RM " $arg"; rmforce=: ;; -*) func_append RM " $arg" ;; *) func_append files " $arg" ;; esac done test -z "$RM" && \ func_fatal_help "you must specify an RM program" rmdirs= for file in $files; do func_dirname "$file" "" "." dir=$func_dirname_result if test . = "$dir"; then odir=$objdir else odir=$dir/$objdir fi func_basename "$file" name=$func_basename_result test uninstall = "$opt_mode" && odir=$dir # Remember odir for removal later, being careful to avoid duplicates if test clean = "$opt_mode"; then case " $rmdirs " in *" $odir "*) ;; *) func_append rmdirs " $odir" ;; esac fi # Don't error if the file doesn't exist and rm -f was used. if { test -L "$file"; } >/dev/null 2>&1 || { test -h "$file"; } >/dev/null 2>&1 || test -f "$file"; then : elif test -d "$file"; then exit_status=1 continue elif $rmforce; then continue fi rmfiles=$file case $name in *.la) # Possibly a libtool archive, so verify it. if func_lalib_p "$file"; then func_source $dir/$name # Delete the libtool libraries and symlinks. for n in $library_names; do func_append rmfiles " $odir/$n" done test -n "$old_library" && func_append rmfiles " $odir/$old_library" case $opt_mode in clean) case " $library_names " in *" $dlname "*) ;; *) test -n "$dlname" && func_append rmfiles " $odir/$dlname" ;; esac test -n "$libdir" && func_append rmfiles " $odir/$name $odir/${name}i" ;; uninstall) if test -n "$library_names"; then # Do each command in the postuninstall commands. func_execute_cmds "$postuninstall_cmds" '$rmforce || exit_status=1' fi if test -n "$old_library"; then # Do each command in the old_postuninstall commands. func_execute_cmds "$old_postuninstall_cmds" '$rmforce || exit_status=1' fi # FIXME: should reinstall the best remaining shared library. ;; esac fi ;; *.lo) # Possibly a libtool object, so verify it. if func_lalib_p "$file"; then # Read the .lo file func_source $dir/$name # Add PIC object to the list of files to remove. if test -n "$pic_object" && test none != "$pic_object"; then func_append rmfiles " $dir/$pic_object" fi # Add non-PIC object to the list of files to remove. if test -n "$non_pic_object" && test none != "$non_pic_object"; then func_append rmfiles " $dir/$non_pic_object" fi fi ;; *) if test clean = "$opt_mode"; then noexename=$name case $file in *.exe) func_stripname '' '.exe' "$file" file=$func_stripname_result func_stripname '' '.exe' "$name" noexename=$func_stripname_result # $file with .exe has already been added to rmfiles, # add $file without .exe func_append rmfiles " $file" ;; esac # Do a test to see if this is a libtool program. if func_ltwrapper_p "$file"; then if func_ltwrapper_executable_p "$file"; then func_ltwrapper_scriptname "$file" relink_command= func_source $func_ltwrapper_scriptname_result func_append rmfiles " $func_ltwrapper_scriptname_result" else relink_command= func_source $dir/$noexename fi # note $name still contains .exe if it was in $file originally # as does the version of $file that was added into $rmfiles func_append rmfiles " $odir/$name $odir/${name}S.$objext" if test yes = "$fast_install" && test -n "$relink_command"; then func_append rmfiles " $odir/lt-$name" fi if test "X$noexename" != "X$name"; then func_append rmfiles " $odir/lt-$noexename.c" fi fi fi ;; esac func_show_eval "$RM $rmfiles" 'exit_status=1' done # Try to remove the $objdir's in the directories where we deleted files for dir in $rmdirs; do if test -d "$dir"; then func_show_eval "rmdir $dir >/dev/null 2>&1" fi done exit $exit_status } if test uninstall = "$opt_mode" || test clean = "$opt_mode"; then func_mode_uninstall ${1+"$@"} fi test -z "$opt_mode" && { help=$generic_help func_fatal_help "you must specify a MODE" } test -z "$exec_cmd" && \ func_fatal_help "invalid operation mode '$opt_mode'" if test -n "$exec_cmd"; then eval exec "$exec_cmd" exit $EXIT_FAILURE fi exit $exit_status # The TAGs below are defined such that we never get into a situation # where we disable both kinds of libraries. Given conflicting # choices, we go for a static library, that is the most portable, # since we can't tell whether shared libraries were disabled because # the user asked for that or because the platform doesn't support # them. This is particularly important on AIX, because we don't # support having both static and shared libraries enabled at the same # time on that platform, so we default to a shared-only configuration. # If a disable-shared tag is given, we'll fallback to a static-only # configuration. But we'll never go from static-only to shared-only. # ### BEGIN LIBTOOL TAG CONFIG: disable-shared build_libtool_libs=no build_old_libs=yes # ### END LIBTOOL TAG CONFIG: disable-shared # ### BEGIN LIBTOOL TAG CONFIG: disable-static build_old_libs=`case $build_libtool_libs in yes) echo no;; *) echo yes;; esac` # ### END LIBTOOL TAG CONFIG: disable-static # Local Variables: # mode:shell-script # sh-indentation:2 # End:
<reponame>leiteszeke/cookunity-ui import Button from './Button'; import { ButtonProps, ButtonIconPosition, ButtonSize, ButtonVariant, } from './Button.types'; export { ButtonProps, ButtonIconPosition, ButtonSize, ButtonVariant }; export default Button;
package com.ibm.socialcrm.notesintegration.ui.utils; /**************************************************************** * IBM OpenSource * * (C) Copyright IBM Corp. 2012 * * Licensed under the Apache License v2.0 * http://www.apache.org/licenses/LICENSE-2.0 * ***************************************************************/ import org.eclipse.core.runtime.NullProgressMonitor; import com.ibm.socialcrm.notesintegration.ui.dashboardcomposites.AbstractDashboardComposite; import com.ibm.socialcrm.notesintegration.utils.ConstantStrings; /** * There are a number of places in the cards where we have internal links to other sugar objects. If those objects aren't already loaded, we need to call the sugar webservices to get them. This can * take a non-trivial amount of time and we want to alert the user that something is going on. * * This is a very specialized progress monitor design to work with an AbstractDashboardComposite. */ public class WebserviceDataLoadProgressMonitor extends NullProgressMonitor { private String progressId; private String message; private AbstractDashboardComposite dashboardComposite; public WebserviceDataLoadProgressMonitor(AbstractDashboardComposite dashboardComposite, String message) { setDashboardComposite(dashboardComposite); setMessage(message); } @Override public void beginTask(String name, int totalWork) { if (getDashboardComposite() != null) { String id = getDashboardComposite().getProgressDisplayer().createProgressIndicator(getMessage()); setProgressId(id); } } @Override public void done() { if (getDashboardComposite() != null) { getDashboardComposite().getProgressDisplayer().removeProgressIndicator(getProgressId()); } } private String getProgressId() { if (progressId == null) { progressId = ConstantStrings.EMPTY_STRING; } return progressId; } private void setProgressId(String progressId) { this.progressId = progressId; } public String getMessage() { if (message == null) { message = ConstantStrings.EMPTY_STRING; } return message; } public void setMessage(String message) { this.message = message; } public AbstractDashboardComposite getDashboardComposite() { return dashboardComposite; } public void setDashboardComposite(AbstractDashboardComposite dashboardComposite) { this.dashboardComposite = dashboardComposite; } }
<reponame>smagill/opensphere-desktop<gh_stars>10-100 package io.opensphere.core.geometry.constraint; import io.opensphere.core.model.time.TimeSpan; import net.jcip.annotations.Immutable; /** * A strict time constraint that requires that the active time span exactly * matches (or doesn't match) a certain time span. */ @Immutable public class StrictTimeConstraint extends TimeConstraint { /** * Constructor. * * @param key The key associated with this constraint. The key must be an * immutable object. * @param startTime The start time. * @param endTime The end time. */ public StrictTimeConstraint(Object key, long startTime, long endTime) { super(key, startTime, endTime); } /** * Constructor. * * @param key The key associated with this constraint. The key must be an * immutable object. * @param startTime The start time. * @param endTime The end time. * @param negate If {@code true}, the check is negated. */ public StrictTimeConstraint(Object key, long startTime, long endTime, boolean negate) { super(key, startTime, endTime, negate, false); } /** * Constructor. * * @param key The key associated with this constraint. The key must be an * immutable object. * @param span The time span that must be covered. */ public StrictTimeConstraint(Object key, TimeSpan span) { super(key, span, false); } /** * Constructor. * * @param key The key associated with this constraint. The key must be an * immutable object. * @param span The time span that must be covered. * @param negate If {@code true}, the check is negated. */ public StrictTimeConstraint(Object key, TimeSpan span, boolean negate) { super(key, span, negate); } @Override public boolean check(TimeSpan span) { return isNegative() ^ getTimeSpan().equals(span); } }
#!/usr/bin/env bash if [[ $TRAVIS_BRANCH == 'master' ]] && [ "$TRAVIS_PULL_REQUEST" = "false" ]; then source travis/extract.sh source travis/docker.sh sbt "+ test" "mleap-serving/test" "+ publishSigned" "mleap-serving/docker:publish" else sbt "+ test" "mleap-serving/test" fi
/** @noSelfInFile */ declare function SupportItemCooldownReset(killedUnit: CBaseEntity, killerEntity: CBaseEntity): void; declare function ForeheadProtectorOnItemPickedUp(hero: CDOTA_BaseNPC_Hero, itemName: string): void; declare function ChakraArmorOnItemPickedUp(hero: CDOTA_BaseNPC_Hero, itemName: string): void;
'use strict'; const config = require('./config'); const sqlite3 = require('sqlite3').verbose(); const db = new sqlite3.Database(config.database_path); db.serialize(function () { db.run('CREATE TABLE games (name TEXT, token TEXT)'); db.run('CREATE INDEX index_by_name on games(name)'); db.run('CREATE TABLE scores (point INTEGER, game_name TEXT, player_name TEXT)'); db.run('CREATE INDEX index_by_game_name on scores(game_name)'); }); db.close();
<filename>src/main/resources/schema.sql -- I'm sorry I had to do this barbaric thing. R2DBC does not support query derivation as of Dec 2019. 😿 CREATE TABLE FETCHLIN_PAGE ( id SERIAL PRIMARY KEY, url_ VARCHAR(255), name_ VARCHAR(255), interval_ INT, max_number_of_revisions INT, dom_element VARCHAR(255), last_fetch_time VARCHAR(50) ); CREATE TABLE FETCHLIN_REVISION ( id SERIAL PRIMARY KEY, data_ VARCHAR(2147483647), fetch_time VARCHAR(50), page_id INT );
import React from "react"; const formsRoutes = [ // // // // // // // project progressive { path: "/projects/oil-gaz-en-cour/", component: React.lazy(() => import("./projectsProgress/oilGaz/OilGaz")) }, { path: "/projects/enrgies-renewable-en-cour", component: React.lazy(() => import("./projectsProgress/RenewableEnergies/RenewableEnergies")) }, { path: "/projects/operation-maintenace-en-cour", component: React.lazy(() => import("./projectsProgress/OperationMaintenance/OperationMaintenance")) }, { path: "/projects/water-trairtement-en-cour", component: React.lazy(() => import("./projectsProgress/WaterTreatment/WaterTreatment")) }, { path: "/projects/industrie", component: React.lazy(() => import("./projectsProgress/industry/Industry")) }, // // // // // // // project progressive show { path: "/oil-gaz-en-cour/:id", component: React.lazy(() => import("./projectsProgress/oilGaz/OilGazShow")) }, { path: "/water-trairtement-en-cour/:id", component: React.lazy(() => import("./projectsProgress/WaterTreatment/ShowWaterTreatment")) }, { path: "/enrgies-renewable-en-cour/:id", component: React.lazy(() => import("./projectsProgress/RenewableEnergies/ShowRenewableEnergies")) }, { path: "/operation-maintenace-en-cour/:id", component: React.lazy(() => import("./projectsProgress/OperationMaintenance/ShowOperationMaintenace")) }, { path: "/industrie/:id", component: React.lazy(() => import("./projectsProgress/industry/ShowIndustry")) }, // // // // // //projets achevés { path: "/projects/oil-gaz-acheve", component: React.lazy(() => import("./completedProject/oilGaz/OilGazCompleted")) }, { path: "/projects/industry-acheve", component: React.lazy(() => import("./completedProject/industry/IndustryComplete")) }, { path: "/projects/operation-acheve", component: React.lazy(() => import("./completedProject/OperationMaintenance/OperationMaintenanceComplete")) }, { path: "/projects/energy-acheve", component: React.lazy(() => import("./completedProject/RenewableEnergies/RenewableEnergiesComplete")) }, { path: "/projects/water-trairtement-acheve", component: React.lazy(() => import("./completedProject/WaterTreatment/WaterTreatmentComplete")) }, // // // // // // // // project show achevés { path: "/oil-gaz-acheve/:id", component: React.lazy(() => import("./completedProject/oilGaz/OilGazCompleteShow")) }, { path: "/industry-acheve/:id", component: React.lazy(() => import("./completedProject/industry/IndustryCompleteShow")) }, { path: "/operation-acheve/:id", component: React.lazy(() => import("./completedProject/OperationMaintenance/OperationMaintenaceCompleteShow")) }, { path: "/energy-acheve/:id", component: React.lazy(() => import("./completedProject/RenewableEnergies/RenewableEnergiesCompleteShow")) }, { path: "/water-trairtement-acheve/:id", component: React.lazy(() => import("./completedProject/WaterTreatment/WaterTreatmentCompleteShow")) }, // // // // // // // // // product { path: "/product", component: React.lazy(() => import("./ProductService/ProductService")) }, // // // // // // // // // Setting { path: "/projects/add-project", component: React.lazy(() => import("./addProjects/AddProjects")) }, { path: "/projects/manage-project", component: React.lazy(() => import("./projectManage/ProjectManage")) }, ]; export default formsRoutes;
package chylex.hee.item.base; import java.util.List; import net.minecraft.client.renderer.texture.IIconRegister; import net.minecraft.creativetab.CreativeTabs; import net.minecraft.entity.Entity; import net.minecraft.entity.player.EntityPlayer; import net.minecraft.init.Items; import net.minecraft.item.EnumAction; import net.minecraft.item.Item; import net.minecraft.item.ItemStack; import net.minecraft.util.IIcon; import net.minecraft.util.StatCollector; import net.minecraft.world.World; import chylex.hee.entity.projectile.EntityProjectilePotion; import cpw.mods.fml.relauncher.Side; import cpw.mods.fml.relauncher.SideOnly; public abstract class ItemAbstractPotion extends Item{ @SideOnly(Side.CLIENT) private IIcon iconBottleNormal, iconBottleSplash; public ItemAbstractPotion(){ setMaxStackSize(1); } public abstract void applyEffectDrunk(ItemStack is, World world, EntityPlayer player); public abstract void applyEffectThrown(Entity entity, double dist); @Override public final ItemStack onEaten(ItemStack is, World world, EntityPlayer player){ if (!world.isRemote)applyEffectDrunk(is, world, player); if (!player.capabilities.isCreativeMode){ if (--is.stackSize <= 0)return new ItemStack(Items.glass_bottle); player.inventory.addItemStackToInventory(new ItemStack(Items.glass_bottle)); } return is; } @Override public final ItemStack onItemRightClick(ItemStack is, World world, EntityPlayer player){ if (is.getItemDamage() == 1){ if (!player.capabilities.isCreativeMode)--is.stackSize; world.playSoundAtEntity(player, "random.bow", 0.5F, 0.4F/(itemRand.nextFloat()*0.4F+0.8F)); if (!world.isRemote)world.spawnEntityInWorld(new EntityProjectilePotion(world, player, this)); return is; } player.setItemInUse(is, getMaxItemUseDuration(is)); return is; } @Override public final int getMaxItemUseDuration(ItemStack is){ return 32; } @Override public final EnumAction getItemUseAction(ItemStack is){ return EnumAction.drink; } @Override public final String getItemStackDisplayName(ItemStack is){ return is.getItemDamage() != 1 ? StatCollector.translateToLocal(getUnlocalizedName()+".name") : StatCollector.translateToLocal(getUnlocalizedName()+".splash.name"); } @Override @SideOnly(Side.CLIENT) public final boolean hasEffect(ItemStack is, int pass){ return pass == 0 && is.getItemDamage() < 2; } @Override @SideOnly(Side.CLIENT) public final boolean requiresMultipleRenderPasses(){ return true; } @Override @SideOnly(Side.CLIENT) public final IIcon getIconFromDamageForRenderPass(int damage, int pass){ return pass == 0 ? itemIcon : damage == 1 ? iconBottleSplash : iconBottleNormal; } @Override @SideOnly(Side.CLIENT) public final void getSubItems(Item item, CreativeTabs tab, List list){ list.add(new ItemStack(item, 1, 0)); list.add(new ItemStack(item, 1, 1)); } @Override @SideOnly(Side.CLIENT) public final void registerIcons(IIconRegister iconRegister){ super.registerIcons(iconRegister); iconBottleNormal = iconRegister.registerIcon("potion_bottle_drinkable"); iconBottleSplash = iconRegister.registerIcon("potion_bottle_splash"); } }
package patron.mains.managers.app; import java.io.File; import java.io.IOException; import java.io.InputStream; import java.net.URLDecoder; import java.util.Properties; /** * The type App manager configuration. */ public class AppManagerConfiguration { private Properties properties; private String resourceIcon; private File profileSettingsFile; private File appSettingsFile; private Long maxTimeClose; private File appDir; private File dataDir; /** * Instantiates a new App manager configuration. * * @throws IOException the io exception */ public AppManagerConfiguration() throws IOException { this.properties = new Properties(); this.appDir = new File(URLDecoder.decode(AppManagerConfiguration.class.getProtectionDomain().getCodeSource().getLocation().getPath(), "UTF-8")); if (this.appDir.isFile()){ this.appDir = this.appDir.getParentFile(); } this.resourceIcon = "/patron/mains/guis/main/images/MainIcon.png"; this.profileSettingsFile = new File(System.getProperty("user.home"), "appscharles/apps/stocker/profileSettings.properties"); this.appSettingsFile = new File( this.appDir , "Settings/app.properties"); InputStream input = getClass().getClassLoader().getResourceAsStream("patron/mains/managers/app/AppManagerConfiguration.properties"); this.properties.load(input); this.maxTimeClose = 60000L; this.dataDir = new File(this.appDir, "Data"); } /** * Gets properties. * * @return the properties */ public Properties getProperties() { return properties; } /** * Sets properties. * * @param properties the properties */ public void setProperties(Properties properties) { this.properties = properties; } /** * Gets resource icon. * * @return the resource icon */ public String getResourceIcon() { return resourceIcon; } /** * Sets resource icon. * * @param resourceIcon the resource icon */ public void setResourceIcon(String resourceIcon) { this.resourceIcon = resourceIcon; } /** * Gets profile settings file. * * @return the profile settings file */ public File getProfileSettingsFile() { return profileSettingsFile; } /** * Sets profile settings file. * * @param profileSettingsFile the profile settings file */ public void setProfileSettingsFile(File profileSettingsFile) { this.profileSettingsFile = profileSettingsFile; } /** * Gets app settings file. * * @return the app settings file */ public File getAppSettingsFile() { return appSettingsFile; } /** * Sets app settings file. * * @param appSettingsFile the app settings file */ public void setAppSettingsFile(File appSettingsFile) { this.appSettingsFile = appSettingsFile; } /** * Gets max time close. * * @return the max time close */ public Long getMaxTimeClose() { return maxTimeClose; } /** * Sets max time close. * * @param maxTimeClose the max time close */ public void setMaxTimeClose(Long maxTimeClose) { this.maxTimeClose = maxTimeClose; } /** * Gets app dir. * * @return the app dir */ public File getAppDir() { return appDir; } /** * Sets app dir. * * @param appDir the app dir */ public void setAppDir(File appDir) { this.appDir = appDir; } /** * Gets data dir. * * @return the data dir */ public File getDataDir() { return dataDir; } /** * Sets data dir. * * @param dataDir the data dir */ public void setDataDir(File dataDir) { this.dataDir = dataDir; } }